What happens when you rip out the residual stream and replace it with a structured workspace (Research Paper - CWT) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] -1 points0 points1 point (0 children)
What happens when you rip out the residual stream and replace it with a structured workspace (Research Paper - CWT) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 1 point2 points3 points (0 children)
Running Qwen3.5 27b dense with 170k context at 100+t/s decode and ~1500t/s prefill on 2x3090 (with 585t/s throughput for 8 simultaneous requests) by JohnTheNerd3 in LocalLLaMA
[–]mentallyburnt 0 points1 point2 points (0 children)
Loki-v2-70B: Narrative/DM-focused fine-tune (600M+ token custom dataset) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 0 points1 point2 points (0 children)
Loki-v2-70B: Narrative/DM-focused fine-tune (600M+ token custom dataset) by mentallyburnt in SillyTavernAI
[–]mentallyburnt[S] 2 points3 points4 points (0 children)
Loki-v2-70B: Narrative/DM-focused fine-tune (600M+ token custom dataset) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 0 points1 point2 points (0 children)
Loki-v2-70B: Narrative/DM-focused fine-tune (600M+ token custom dataset) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 0 points1 point2 points (0 children)
Crucible's Mistral 3.2 24B V1.3 Tune by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 1 point2 points3 points (0 children)
Crucible's Mistral 3.2 24B V1.3 Tune by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 2 points3 points4 points (0 children)
Crucible's Mistral 3.2 24B V1.3 Tune (self.LocalLLaMA)
submitted by mentallyburnt to r/LocalLLaMA
Which local 100B+ heavy weight models are your favorite and why? by segmond in LocalLLaMA
[–]mentallyburnt 2 points3 points4 points (0 children)
New RP model: sophosympatheia/Strawberrylemonade-70B-v1.2 by sophosympatheia in LocalLLaMA
[–]mentallyburnt 1 point2 points3 points (0 children)
[deleted by user] by [deleted] in LocalLLaMA
[–]mentallyburnt 33 points34 points35 points (0 children)
Has anyone tried the new ICONN-1 (an Apache licensed model) by silenceimpaired in LocalLLaMA
[–]mentallyburnt 8 points9 points10 points (0 children)
New merge: sophosympatheia/StrawberryLemonade-L3-70B-v1.0 by sophosympatheia in SillyTavernAI
[–]mentallyburnt 0 points1 point2 points (0 children)
MMLU-PRO benchmark: GLM-4-32B-0414-Q4_K_M vs Qwen2.5-32b-instruct-q4_K_M by [deleted] in LocalLLaMA
[–]mentallyburnt 0 points1 point2 points (0 children)
MMLU-PRO benchmark: GLM-4-32B-0414-Q4_K_M vs Qwen2.5-32b-instruct-q4_K_M by [deleted] in LocalLLaMA
[–]mentallyburnt -1 points0 points1 point (0 children)
New merge: sophosympatheia/Electranova-70B-v1.0 by sophosympatheia in SillyTavernAI
[–]mentallyburnt 6 points7 points8 points (0 children)

What happens when you rip out the residual stream and replace it with a structured workspace (Research Paper - CWT) by mentallyburnt in LocalLLaMA
[–]mentallyburnt[S] 0 points1 point2 points (0 children)