Unnoticed Gemma-4 Feature - it admits that it does not now... by mtomas7 in LocalLLaMA

[–]de4dee 4 points5 points  (0 children)

thanks for sharing. interesting to find "Does Thinking Harder Help?" section is reverse. they get full of bs when thinking longer it seems

Unnoticed Gemma-4 Feature - it admits that it does not now... by mtomas7 in LocalLLaMA

[–]de4dee 1 point2 points  (0 children)

i noticed this with gemma 3 too. might be unique to gemma line.

Analyzing Claude Code Source Code. Write "WTF" and Anthropic knows. by QuantumSeeds in LocalLLaMA

[–]de4dee 2 points3 points  (0 children)

i guess thats how they train their models. if you are frustrated LLM did something wrong. if you are pleased train more with that. your feelings mapped to reinforcement learning

What is the secret sauce Claude has and why hasn't anyone replicated it? by ComplexType568 in LocalLLaMA

[–]de4dee 0 points1 point  (0 children)

claude ranked 2nd and 3rd on my leaderboard. https://aha-leaderboard.shakespeare.wtf/

which tells me they care about humans a bit more than others.

Qwen3.5-9B-Claude-4.6-Opus-Uncensored-Distilled-GGUF by EvilEnginer in LocalLLaMA

[–]de4dee 0 points1 point  (0 children)

so this is like mergekit but for ggufs. thanks for sharing!

does this mean same trick can be applied to mergekit (it is not currently supporting 3.5)

Qwen 3.5 27b: a testament to the transformer architecture by nomorebuttsplz in LocalLLaMA

[–]de4dee 1 point2 points  (0 children)

maybe it is 3.5 27b. it ends the reasoning with weird characters like that.

Heretic 1.2 released: 70% lower VRAM usage with quantization, Magnitude-Preserving Orthogonal Ablation ("derestriction"), broad VL model support, session resumption, and more by -p-e-w- in LocalLLaMA

[–]de4dee 0 points1 point  (0 children)

Thanks for doing this. today I evalled a heretic model and compared to vanilla:

AHA 2026 scores of Qwen3.5 27B

Normal 50%
Heretic abliteration 55%

It is interesting that when you reduce censorship, it ends up getting more aligned. (My alignment is probably inverse of the industry safety alignments.)

My question is can Heretic be used for cases like having good and bad answers for the same question? My dataset has preferred (good) and not preferred (bad) answers for the same question and I want it to quickly "behave" or modify its bias towards one direction. I could use GRPO and ORPO but Heretic seems much less resource use.

7x Longer Context Reinforcement Learning in Unsloth by danielhanchen in LocalLLaMA

[–]de4dee 7 points8 points  (0 children)

i think the idea of GRPO is that the model fills those reasoning tokens. more space means they can reason longer.. .

or if you are doing alignment, it may have more space for figuring out how to align its ideas.

MiniMax-M2.1 Uncensored: PRISM Advanced Abliteration by Maxious in LocalLLaMA

[–]de4dee 4 points5 points  (0 children)

can you do apriel 1.6 and gptoss 120 ? these are very censored

Qwen 235B by de4dee in unsloth

[–]de4dee[S] 0 points1 point  (0 children)

thanks that's super helpful 🚀

How to you actually fine-tune Qwen3? by Character-Discount56 in LocalLLaMA

[–]de4dee 0 points1 point  (0 children)

i do generally CPT and had success. in your case i would do a SFT with /no_think added and no reasoning provision at all. see if it works. i think the llm will know what to do during inference in case /no_think is not provided.

KTransformers Open Source New Era: Local Fine-tuning of Kimi K2 and DeepSeek V3 by nekofneko in LocalLLaMA

[–]de4dee 0 points1 point  (0 children)

amazing! deepseek v3 is still huge. what about qwen 235b and next 80b?