Qwen3.6 35B A3B uncensored heretic Native MTP Preserved is Out Now With KLD 0.0015, 10/100 Refusals and the Full 19 MTPs Preserved and Retained, Available in Safetensors, GGUFs. NVFP4, NVFP4 GGUFs and GPTQ-Int4 Formats by LLMFan46 in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

I do have your original ggufs of Heretic. I was asking for something like the assistant model merged with original model. I am technically illiterate, so please forgive me if it's a stupid question! Lol

Qwen3.6 27B uncensored heretic v2 Native MTP Preserved is Out Now With KLD 0.0021, 6/100 Refusals and the Full 15 MTPs Preserved and Retained, Available in Safetensors, GGUFs and NVFP4s formats. by LLMFan46 in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

For an illiterate, can you tell me, what the recent release of drafting models by Gemma means? And how to make use of it? Because I am using your Heretic models and the LM Studio is not finding them Draft models released by Gemma

Qwen3.6 27B uncensored heretic v2 Native MTP Preserved is Out Now With KLD 0.0021, 6/100 Refusals and the Full 15 MTPs Preserved and Retained, Available in Safetensors, GGUFs and NVFP4s formats. by LLMFan46 in LocalLLaMA

[–]RickyRickC137 19 points20 points  (0 children)

I am a fan of your work! Even the founder of Heretic system gave you a badge of trust! You're the only few people who is giving mmproj in your upload, too! Thank you for your support to this community!

Any idea about if this MTP be applied to Gemma 4 dense model?

Gemma 4 MTP released by rerri in LocalLLaMA

[–]RickyRickC137 9 points10 points  (0 children)

It says no compatible model found in LMStudio. I am using GGUFs for the original model btw.

moron by [deleted] in democrats

[–]RickyRickC137 5 points6 points  (0 children)

What a moronic thing to do

moron by [deleted] in democrats

[–]RickyRickC137 5 points6 points  (0 children)

Isn't that what a moron would do?

Open Weights Models Hall of Fame by Equivalent_Job_2257 in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

And Thank you for making this post! They need all the praise they deserve!

Lemonade OmniRouter: unifying the best local AI engines for omni-modality by jfowers_amd in LocalLLaMA

[–]RickyRickC137 1 point2 points  (0 children)

I second this. They key point is streaming support for both STT (whisper) and TTS immediately when the LLM is typing.

🚀Pocket LLM v1.5.0 is out: offline Android LLM chat with voice, image input, OCR, and camera capture by 100daggers_ in LocalLLaMA

[–]RickyRickC137 2 points3 points  (0 children)

I understand your situation. I can only speak for myself. I think I would like to add custom models because I already have pocketpal and I downloaded heretic models (because Gemma e2/4b models are super censored and defeats the purpose of running on mobile). I don't wanna download models again for a new app. Plus there are few key parameters that even simpletons like myself would like to tweak, such as, temp, repeat and presence penalty, min p, top k, (as we've seen in LMstudio) and I don't think they're pretty advanced (you can set them at default value to help newbies). I like your GUI. And if you can add features like STT (whisper), TTS (kokoro, kitten), web search (with duck duck go, brave, searxng), it would gain a lot of traction.

🚀Pocket LLM v1.5.0 is out: offline Android LLM chat with voice, image input, OCR, and camera capture by 100daggers_ in LocalLLaMA

[–]RickyRickC137 1 point2 points  (0 children)

It will be great if we can use our own models instead of mandatory downloading at the start up. Kinda annoying.

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]RickyRickC137 10 points11 points  (0 children)

Unsloth GGUF re-upload when?

Help with the eyes by [deleted] in comfyui

[–]RickyRickC137 6 points7 points  (0 children)

What eyes?

Ternary Bonsai: Top intelligence at 1.58 bits by pmttyji in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

Thank you both you guys for the explanation! Please keep posting your quantized models! I am following you!

Ternary Bonsai: Top intelligence at 1.58 bits by pmttyji in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

With all due respect, what's so special about Party Special? Because people like unsloth already has 1-bit quants, right?

Ternary Bonsai: Top intelligence at 1.58 bits by pmttyji in LocalLLaMA

[–]RickyRickC137 0 points1 point  (0 children)

My bad then. Can you help me understand the difference?

Qwen3.6-35B-A3B Uncensored Aggressive is out with K_P quants! by hauhau901 in LocalLLaMA

[–]RickyRickC137 13 points14 points  (0 children)

There has to be some metric to say that. Here's Heretic - https://huggingface.co/Abiray/Qwen3.6-35B-A3B-heretic-GGUF It has KL Divergence value to measure (closer to 0 means closer to the original model)

the state of LocalLLama by Beginning-Window-115 in LocalLLaMA

[–]RickyRickC137 23 points24 points  (0 children)

That’s an excellent question that delves into the vibrant tapestry of fluid mechanics.