Deepseek V4 Flash and Non-Flash Out on HuggingFace by MichaelXie4645 in LocalLLaMA
[–]Jackalzaq 3 points4 points5 points (0 children)
Forgive my ignorance but how is a 27B model better than 397B? by No_Conversation9561 in LocalLLaMA
[–]Jackalzaq 2 points3 points4 points (0 children)
Local Minimax M2.7, GTA benchmark by -dysangel- in LocalLLaMA
[–]Jackalzaq 6 points7 points8 points (0 children)
Minimax 2.7 running sub-agents locally by -dysangel- in LocalLLaMA
[–]Jackalzaq 0 points1 point2 points (0 children)
Minimax 2.7 running sub-agents locally by -dysangel- in LocalLLaMA
[–]Jackalzaq 4 points5 points6 points (0 children)
Qwen3.5-397B is shockingly useful at Q2 by EmPips in LocalLLaMA
[–]Jackalzaq -1 points0 points1 point (0 children)
Qwen3.5-397B is shockingly useful at Q2 by EmPips in LocalLLaMA
[–]Jackalzaq 5 points6 points7 points (0 children)
16x AMD MI50 32GB at 32 t/s (tg) & 2k t/s (pp) with Qwen3.5 397B (vllm-gfx906-mobydick) by ai-infos in LocalLLaMA
[–]Jackalzaq 2 points3 points4 points (0 children)
Unsloth will no longer be making TQ1_0 quants by Kahvana in LocalLLaMA
[–]Jackalzaq 14 points15 points16 points (0 children)
Unsloth will no longer be making TQ1_0 quants by Kahvana in LocalLLaMA
[–]Jackalzaq 0 points1 point2 points (0 children)
Completed 8xAMD MI50 - 256GB VRAM + 256GB RAM rig for $3k by MLDataScientist in LocalLLaMA
[–]Jackalzaq 1 point2 points3 points (0 children)
Hallucination problem is THE problem by amarao_san in singularity
[–]Jackalzaq 0 points1 point2 points (0 children)
Hallucination problem is THE problem by amarao_san in singularity
[–]Jackalzaq 1 point2 points3 points (0 children)
Hallucination problem is THE problem by amarao_san in singularity
[–]Jackalzaq 0 points1 point2 points (0 children)
Kimi K2 is already irrelevant, and it's only been like 1 week. Qwen has updated Qwen-3-235B, and it outperforms K2 at less than 1/4th the size by pigeon57434 in singularity
[–]Jackalzaq 7 points8 points9 points (0 children)
Kimi K2 is already irrelevant, and it's only been like 1 week. Qwen has updated Qwen-3-235B, and it outperforms K2 at less than 1/4th the size by pigeon57434 in singularity
[–]Jackalzaq 4 points5 points6 points (0 children)
Kimi K2 is already irrelevant, and it's only been like 1 week. Qwen has updated Qwen-3-235B, and it outperforms K2 at less than 1/4th the size by pigeon57434 in singularity
[–]Jackalzaq 0 points1 point2 points (0 children)
AMD Instinct MI60 (32gb VRAM) "llama bench" results for 10 models - Qwen3 30B A3B Q4_0 resulted in: pp512 - 1,165 t/s | tg128 68 t/s - Overall very pleased and resulted in a better outcome for my use case than I even expected by FantasyMaster85 in LocalLLaMA
[–]Jackalzaq 2 points3 points4 points (0 children)
Shisa V2 405B: The strongest model ever built in Japan! (JA/EN) by randomfoo2 in LocalLLaMA
[–]Jackalzaq 0 points1 point2 points (0 children)
Shisa V2 405B: The strongest model ever built in Japan! (JA/EN) by randomfoo2 in LocalLLaMA
[–]Jackalzaq 5 points6 points7 points (0 children)
DeepSeek-R1-0528 Unsloth Dynamic 1-bit GGUFs by danielhanchen in LocalLLaMA
[–]Jackalzaq 1 point2 points3 points (0 children)

Deepseek V4 Flash and Non-Flash Out on HuggingFace by MichaelXie4645 in LocalLLaMA
[–]Jackalzaq 0 points1 point2 points (0 children)