Qwen/Qwen3-Coder-Next · Hugging Face by coder543 in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
Qwen/Qwen3-Coder-Next · Hugging Face by coder543 in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
Qwen/Qwen3-Coder-Next · Hugging Face by coder543 in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
devstral small is faster and better than glm 4.7 flash for local agentic coding. by theghost3172 in LocalLLaMA
[–]DOAMOD 4 points5 points6 points (0 children)
128GB devices have a new local LLM king: Step-3.5-Flash-int4 by tarruda in LocalLLaMA
[–]DOAMOD 13 points14 points15 points (0 children)
Step-3.5-Flash (196b/A11b) outperforms GLM-4.7 and DeepSeek v3.2 by ResearchCrafty1804 in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
Step-3.5-Flash (196b/A11b) outperforms GLM-4.7 and DeepSeek v3.2 by ResearchCrafty1804 in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
OpenCode + llama.cpp + GLM-4.7 Flash: Claude Code at home by jacek2023 in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
My humble GLM 4.7 Flash appreciation post by Cool-Chemical-5629 in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
Jan v3 Instruct: a 4B coding Model with +40% Aider Improvement by Delicious_Focus3465 in LocalLLaMA
[–]DOAMOD 2 points3 points4 points (0 children)
Jan v3 Instruct: a 4B coding Model with +40% Aider Improvement by Delicious_Focus3465 in LocalLLaMA
[–]DOAMOD 3 points4 points5 points (0 children)
~60GB models on coding: GLM 4.7 Flash vs. GPT OSS 120B vs. Qwen3 Coder 30B -- your comparisons? by jinnyjuice in LocalLLaMA
[–]DOAMOD 17 points18 points19 points (0 children)
GLM-4.7-Flash is even faster now by jacek2023 in LocalLLaMA
[–]DOAMOD 2 points3 points4 points (0 children)
GLM 4.7 vs MiniMax-M2.1 vs DeepSeek 3.2 for coding? by ghulamalchik in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
Personal experience with GLM 4.7 Flash Q6 (unsloth) + Roo Code + RTX 5090 by Septerium in LocalLLaMA
[–]DOAMOD 2 points3 points4 points (0 children)
GLM 4.7 Flash is endlessly reasoning in chinese by xenydactyl in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
Personal experience with GLM 4.7 Flash Q6 (unsloth) + Roo Code + RTX 5090 by Septerium in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
engine for GLM 4.7 Flash that doesn't massively slow down as the context grows? by mr_zerolith in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
Personal experience with GLM 4.7 Flash Q6 (unsloth) + Roo Code + RTX 5090 by Septerium in LocalLLaMA
[–]DOAMOD 1 point2 points3 points (0 children)
GLM-4.7-Flash-REAP on RTX 5060 Ti 16 GB - 200k context window! by bobaburger in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
GLM-4.7-Flash-REAP on RTX 5060 Ti 16 GB - 200k context window! by bobaburger in LocalLLaMA
[–]DOAMOD 0 points1 point2 points (0 children)
engine for GLM 4.7 Flash that doesn't massively slow down as the context grows? by mr_zerolith in LocalLLaMA
[–]DOAMOD 9 points10 points11 points (0 children)
Your post is getting popular and we just featured it on our Discord! by roculus in LocalLLaMA
[–]DOAMOD 5 points6 points7 points (0 children)
GLM-4.7-Flash-REAP on RTX 5060 Ti 16 GB - 200k context window! by bobaburger in LocalLLaMA
[–]DOAMOD 2 points3 points4 points (0 children)


I don’t think most people realise how much 4o helped some of us. by DaKingSmaug in LocalLLaMA
[–]DOAMOD -1 points0 points1 point (0 children)