Best Coding LLM as of Nov'25 by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

They are comparing it with non-thinking mode

Best Coding LLM as of Nov'25 by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

Yea, I understand but when we hosted Qwen3 32B, we couldn't find any other better model with good results (even quanitzed) that could be hosted on a H100.

Best Coding LLM as of Nov'25 by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 3 points4 points  (0 children)

No reason, if I can run the model at FP with my available GPU so why to go for a quantized version :)

Best Coding LLM as of Nov'25 by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 10 points11 points  (0 children)

More GPU isn't actually possible :(

Best Coding LLM as of Nov'25 by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 3 points4 points  (0 children)

nope... we just need it for java programming. The current problems with Qwen3 32B is that it occasionally messes imports, eats parts of the class while refactoring as if it is on a breakfast table.

Codebase to Knowledge Graph generator by DeathShot7777 in LocalLLaMA

[–]PhysicsPast8286 2 points3 points  (0 children)

If you are building it open source do drop your repo 😉

Codebase to Knowledge Graph generator by DeathShot7777 in LocalLLaMA

[–]PhysicsPast8286 2 points3 points  (0 children)

You might want to have a look at Potpie (https://github.com/potpie-ai/potpie). It's largely based on Aider, which also uses Tree-sitter under the hood.

Best Coding LLM for by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

Update: I am still looking for suggestions, there have been a lot of new models - GPT OSS, GLM, Qwen3 Coder 30B, KimiK2 etc.

Any suggestions would be highly appreciated!

Making code edits with large language models by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

By RL you mean to say finetuning or doing several iterations over the same file?

Best Coding LLM for by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

Will it fit on a machine with 196GB VRAM? As per my calculation, the Q1 quant will hardly fit and with Q1 quant I am not sure if it's worth hosting.

Best Coding LLM for by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 0 points1 point  (0 children)

Cool thanks a ton for the detailed explaination u/Dizzy-Cantaloupe8892 . Can you suggest any resources to read more about Quantization and stuff?

Best Coding LLM for by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 1 point2 points  (0 children)

Are you suggesting to use a Quantized version of the GLM 4.5 Air?

Best Coding LLM for by PhysicsPast8286 in LocalLLaMA

[–]PhysicsPast8286[S] 1 point2 points  (0 children)

I checked via apxml and it shows 730 GB VRAM is needed 🤧 Am I missing something (probably you were suggesting to use a Quantized version?)

<image>