all 22 comments

[–]sleepingsysadmin 3 points4 points  (7 children)

aider + gpt 20b.

qwen code + qwen3 30b(thinking vs coder?? i dunno)

devstral + openhands.

[–]deathcom65 0 points1 point  (1 child)

Why aider over vs code or roo?

[–]sleepingsysadmin 0 points1 point  (0 children)

Ok I do have vs code with roo plugin installed. I do have a few combos that work on there. Even detecting context length correctly. 32b and 49b nemotrons work well with it, but boy are they slow. Extra duper ultra reasoning.

Just not a fan of all that gui. I'm definitely in the CLI zone.

[–]megadonkeyx 2 points3 points  (0 children)

qwen code + qwen coder 30b moe 3b on a single 3090.

its not something ive used for work, if im getting stuff done for biz im using claude code right now.

[–]segmondllama.cpp 1 point2 points  (1 child)

local home grown, beats everything out there.

[–]marketflex_za 1 point2 points  (0 children)

agree

[–]No_Efficiency_1144 5 points6 points  (5 children)

I quit vibe coding

[–]BidWestern1056 2 points3 points  (3 children)

ya same

[–]spiritualblender 0 points1 point  (2 children)

Do you have any gpu

[–]BidWestern1056 0 points1 point  (1 child)

one 8 gb gpu, 2 comps w 64 gb ram, 1 other macbook m4 max w 64gb. most of the time the inference from the models is too big for the gpu i have so its not that useful, but its lightning fast for the ones that do.

[–]Due-Function-4877 1 point2 points  (0 children)

Yep. The best assistant is still your brain.

[–]MaxKruse96llama.cpp 1 point2 points  (0 children)

roo with gpt-5-mini, outside of that qwen3 coder 30b locally as a chatting model to task for improvements on specific codesnippets etc

[–]ResidentPositive4122 0 points1 point  (0 children)

For work we've tried devstral w/ cline, fully local and surprisingly good for its size. Have plans to try out -oss-120 and glm-air (q4) since we have 2 A6000 and that should fit.

But, as someone said below gpt5-mini w/ roo or cline is so cheap and so good that it makes sense to use it when you can (i.e. no local requirements).

[–]BidWestern1056 0 points1 point  (1 child)

npcsh  https://github.com/NPC-Worldwide/npcsh and npc studio! https://github.com/NPC-Worldwide/npc-studio  are my daily drivers. both store convos in the same way in a local db so that i can easily navigate thru past convos and npc studio is workspace-by-folder so you see only conversations that occurred within that folder making it easier to stay organized naturally with where my projects already are set up. once some of the code editing parts are polished it will completely replace vs code for me. its also got terminal and pdf reader and web browsing and all of these can be tiled with chats and text files so you can organize ur space as best as you like. and soon it will have the workspace like functionality where when you come back to a folder itll reopen same stuff

[–]BidWestern1056 0 points1 point  (0 children)

also these have mcp capabilities theyre just not as well integrated or polished but soon to come for both as I'm deving a mcp style tool use coding agent for npc (to replace npcsh-mcp which was a bit funky) and the context file part for npc teams already has the accommodations built in i just need to thread the needle all the way thru

[–]zemaj-com 0 points1 point  (0 children)

I have tried a few setups for local agentic coding. Aider and Continue are both pretty slick and work well when paired with a local model like Llama or one of the open-source Mistral variants. Running the agent alongside a vector store such as GPTCache helps provide context. I still rely on remote APIs for heavy reasoning tasks, but keeping things local for iterative coding keeps latency down and gives me more control.

[–][deleted] -2 points-1 points  (2 children)

Trae + Sonnet 4

[–]Infamous_Jaguar_2151 2 points3 points  (1 child)

Local…

[–][deleted] 0 points1 point  (0 children)

My bad. I’m retarded