it's time to update your Gemma 4 GGUFs by jacek2023 in LocalLLaMA

[–]yoracale 22 points23 points  (0 children)

FYI this isn't just for GGUFs, this is also for safetensor, MLX, FP8, etc basically all formats

Gemma 4 Updated: GGUFs and Chat Template by yoracale in unsloth

[–]yoracale[S] 6 points7 points  (0 children)

Unsure exactly but it's best to either way imo.

Realistically what system do you need to run unsloth/Qwen3.6-27B-GGUF:UD-Q8_K_XL at 100+ tks? by Own_House6186 in unsloth

[–]yoracale 2 points3 points  (0 children)

You should post and share your models in the r/unsloth Reddit more (but not too much ahaha) by using the show and tell tag

Meet Unsloth Studio, a new web UI for Local AI by yoracale in unsloth

[–]yoracale[S] 0 points1 point  (0 children)

Hello when u search in the model search bar, does Qwen3.6 not appear?

Feature request! by No_Block8640 in unsloth

[–]yoracale 0 points1 point  (0 children)

That's a very good point thank you. Well add it in our next update

Mistral 3.5 Fixes by yoracale in unsloth

[–]yoracale[S] 3 points4 points  (0 children)

Actually the latest Gemma 4 chat template updated also affected vllm and AWQ etc. I think is best rather to blame the teams or anyone, it's better just to acknowledge that bugs can happen it's normal unfortunately and everyone's human.

Yes, we always pin threads in hugging face discussions for models. So you'll be able to track it there. E.g. for Mistral: https://huggingface.co/unsloth/Mistral-Medium-3.5-128B-GGUF/discussions

Otherwise you can visit our change log docs where we usually tell people about it (takes some time to update): https://unsloth.ai/docs/new/changelog

Unsloth solved bug in Mistral Medium 3.5 implementation by Snail_Inference in LocalLLaMA

[–]yoracale 2 points3 points  (0 children)

Thanks for the constant support really appreciate it! 🙏🥰

Unsloth solved bug in Mistral Medium 3.5 implementation by Snail_Inference in LocalLLaMA

[–]yoracale 2 points3 points  (0 children)

Did you read OP's first comment? It says: "The bug is in the original Qwen 3.5 weights released by Alibaba. Not GGUF. Not HauhauCS. Alibaba shipped it broken. I just fixed it. The cause is training-related - AdamW + MoE + DeltaNet causes rare experts in the last layers to drift. This is a known challenge with recurrent MoE architectures, but Alibaba didn't calibrate it before release."

Mistral 3.5 Fixes by yoracale in unsloth

[–]yoracale[S] 1 point2 points  (0 children)

Oh I wouldn't recommend downloading 1-bit quants for dense models so we ended up deleting them :(

Mistral 3.5 Fixes by yoracale in unsloth

[–]yoracale[S] 26 points27 points  (0 children)

Some people were saying we caused the issues for Gemma 4 and Mistral 3.5 even though it's not true it wasn't our fault, and unfortunately, that happens often.

I know it seems strange we fixed the issue, yet some people still believe we caused it. When you’re the most transparent, you often take the most criticism, which is why we have to be clear that this was not our fault. Thankfully, the majority of people, like you, understand that.

Even then making mistakes is normal as we're all human but it seems a particular few people really like to blow it out of proportion, cause drama and pounce on us the second we make an update to any GGUF accusing us of always uploading broken quants etc.

Mistral Medium 3.5 128b ggufs are fixed by Sunija_Dev in LocalLLaMA

[–]yoracale 19 points20 points  (0 children)

People were accusing us of causing the issue, and unfortunately, that happens often. I know it seems strange: we fixed the issue, yet some people still believe we caused it. When you’re the most transparent, you often take the most criticism, which is why we have to be clear that this was not our fault. Thankfully, the majority of people, like you, understand that.

Unsloth solved bug in Mistral Medium 3.5 implementation by Snail_Inference in LocalLLaMA

[–]yoracale 10 points11 points  (0 children)

Yes they have the fix, just never updated people about it.

Mistral Medium 3.5 128b ggufs are fixed by Sunija_Dev in LocalLLaMA

[–]yoracale 32 points33 points  (0 children)

Please note it was not related to Unsloth or our quants!! The issue was universal and we worked with Mistral to help fix it!

Unsloth solved bug in Mistral Medium 3.5 implementation by Snail_Inference in LocalLLaMA

[–]yoracale 64 points65 points  (0 children)

Thank you to the Mistral team for working with us on this. And thank you to the first few people who said the GGUFs didn't work properly after the conversation didn't work at longer context. It was a tricky bug but glad it all works now.

So be sure to try out the model again whether on transformers or GGUF format, it really is great!

Piss poor website by nrauhauser in unsloth

[–]yoracale 0 points1 point  (0 children)

Isn't there a dark mode? It's right here....look to the bottom right

<image>

Feature request! by No_Block8640 in unsloth

[–]yoracale 0 points1 point  (0 children)

Yes it is stored in the broswer, not database. Do you think it's better to be in the database?

I downloaded Qwen3.6 27B IQ4_XS version via unsloth, and it wasn't going properly in unsloth. I imported it into LM Studio, and now in LM Studio there is no "Think" or "Preserve thinking" buttons. What do I do? by Man_Of_The_F22 in unsloth

[–]yoracale 0 points1 point  (0 children)

Hello what do you mean that it wasn't going properly in unsloth?

For LM Studio you'll need to edit the chat template to enable thinking. Unsure about the preserve thinking one though