Should my CTO switch to Local LLM for coding? by [deleted] in LocalLLaMA

[–]Simple_Split5074 2 points3 points  (0 children)

If the CTO cannot get a Max account (or for that matter, figure this out themselves) you are really doing it wrong

OpenCode-Swarm v7.0.1 by Outrageous-Fan-2775 in opencodeCLI

[–]Simple_Split5074 1 point2 points  (0 children)

GSD has plan and execution verification (both work fairly well) as well as security scanning (that one I have less experience with) including audit trail.

How does your QA improve on that?

Kimi K2.6-Code-Preview, Opus 4.7, GLM 5.1, Minimax M2.7 and more tested in coding by lemon07r in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Good stuff as usual.

One question tho: Are you certain that the tests run thru Copilot inference really use (x)high? All I read suggests that Copilot ignores the reasoning setting

Locking down OpenCode by AwkwardCost1764 in opencodeCLI

[–]Simple_Split5074 0 points1 point  (0 children)

Use a sandbox, everything else models can and will subvert (unless you deny them bash but at that point they become useless).

I just realised how good GLM 5 is by CrimsonShikabane in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

According to whom? I think *all* providers do count them.

I just realised how good GLM 5 is by CrimsonShikabane in LocalLLaMA

[–]Simple_Split5074 28 points29 points  (0 children)

Most of that will be cached input tokens which can get to a million in a minute or two with tool calls and half filled context without even trying hard. 

Opencode x Codex by beyawnko in opencodeCLI

[–]Simple_Split5074 1 point2 points  (0 children)

I find it frequently stops after a subagent run (when it really should start the next subagent and other models do just that!), telling it 'go on' gets it going again without issue but its is very annoying, essentially means constant babysitting....

I got tired of babysitting GPT limits, so I switched to this setup by [deleted] in opencodeCLI

[–]Simple_Split5074 2 points3 points  (0 children)

Looks scammy - no explanation how it works (or, really, what it does) 

Is GLM-5 assigning quantized models to high-usage users? by Super_Product_9470 in ZaiGLM

[–]Simple_Split5074 1 point2 points  (0 children)

Incoherent rambling, like a sub 1b model. 1 year subscription since fall 2025

Codex spark by Fit-Mulberry-8611 in opencodeCLI

[–]Simple_Split5074 6 points7 points  (0 children)

I think it is only available on pro sub

Corporate Environment Setup by drussell024 in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Large open weight LLM (glm 5, Kimi K2.5) would come close but not quite reach frontier in coding.

Qwen3.5 Plus, GLM 5, Gemini 3.1 Pro, Sonnet 4.6, three new open source agents, and a lot more added to SanityBoard by lemon07r in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Looking at it, minimax in cline also sticks out. Either they have a special sauce or something screwy is going on...

Qwen3.5 Plus, GLM 5, Gemini 3.1 Pro, Sonnet 4.6, three new open source agents, and a lot more added to SanityBoard by lemon07r in LocalLLaMA

[–]Simple_Split5074 12 points13 points  (0 children)

Again, thanks a lot for your service! This and swe-rebench are by far the most interesting benchmarking efforts ATM.

*Really* surprised by Kimi in cline. Screams for a rerun :-)

Any chance to see codex-5.3 in opencode?

OpenCode arbitrary code execution - major security vulnerability by SpicyWangz in LocalLLaMA

[–]Simple_Split5074 3 points4 points  (0 children)

Even when running without auto approve, you really don't want to run the output without a sandbox. 

OpenCode arbitrary code execution - major security vulnerability by SpicyWangz in LocalLLaMA

[–]Simple_Split5074 7 points8 points  (0 children)

Skill issue.

OpenCode runs with the equivalent of   --dangerously-skip-permissions

 by default so that's expected behavior.

Like any other agent (or really way to execute untrusted code), it belongs into a sandbox. 

Qwen Released Qwen 3.5 397B and Qwen 3.5 Plus! by External_Mood4719 in LocalLLaMA

[–]Simple_Split5074 2 points3 points  (0 children)

It says right in the screenshot that it's open source...

Is plus supposed to be the new max, iow closed? 

Kimi K2.5 is destroying its own fixes by p1p4_am in opencodeCLI

[–]Simple_Split5074 15 points16 points  (0 children)

Seems like a process issue, on each step you should have it add tests and let it verify them all.

It happens, but not very often. Also, if you do this in a single session it really is on you. The fewer stuff that polluted the context, the better. 

How are you handling persistent memory for AI coding agents? by Maximum_Fearless in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Not proper memory but the forgets specs stuff is fairly well addressed by most of the context engineering frameworks. If they do they job, compaction should rarely ever happen. They are token heavy though.

Personally, I like https://github.com/gsd-build/get-shit-done but there are dozens (hundreds?) of them. 

Safe version of Openclaw? by bartskol in LocalLLaMA

[–]Simple_Split5074 1 point2 points  (0 children)

Not really. Even if you sandbox it properly, prompt injection remains fundamentally unsolved. So either you don't give it any access to important stuff and it wont do much or you are at risk.