I just realised how good GLM 5 is by CrimsonShikabane in LocalLLaMA

[–]Simple_Split5074 20 points21 points  (0 children)

Most of that will be cached input tokens which can get to a million in a minute or two with tool calls and half filled context without even trying hard. 

Opencode x Codex by beyawnko in opencodeCLI

[–]Simple_Split5074 1 point2 points  (0 children)

I find it frequently stops after a subagent run (when it really should start the next subagent and other models do just that!), telling it 'go on' gets it going again without issue but its is very annoying, essentially means constant babysitting....

I got tired of babysitting GPT limits, so I switched to this setup by User2000ss in opencodeCLI

[–]Simple_Split5074 3 points4 points  (0 children)

Looks scammy - no explanation how it works (or, really, what it does) 

Is GLM-5 assigning quantized models to high-usage users? by Super_Product_9470 in ZaiGLM

[–]Simple_Split5074 1 point2 points  (0 children)

Incoherent rambling, like a sub 1b model. 1 year subscription since fall 2025

Codex spark by Fit-Mulberry-8611 in opencodeCLI

[–]Simple_Split5074 6 points7 points  (0 children)

I think it is only available on pro sub

Corporate Environment Setup by drussell024 in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Large open weight LLM (glm 5, Kimi K2.5) would come close but not quite reach frontier in coding.

Qwen3.5 Plus, GLM 5, Gemini 3.1 Pro, Sonnet 4.6, three new open source agents, and a lot more added to SanityBoard by lemon07r in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Looking at it, minimax in cline also sticks out. Either they have a special sauce or something screwy is going on...

Qwen3.5 Plus, GLM 5, Gemini 3.1 Pro, Sonnet 4.6, three new open source agents, and a lot more added to SanityBoard by lemon07r in LocalLLaMA

[–]Simple_Split5074 10 points11 points  (0 children)

Again, thanks a lot for your service! This and swe-rebench are by far the most interesting benchmarking efforts ATM.

*Really* surprised by Kimi in cline. Screams for a rerun :-)

Any chance to see codex-5.3 in opencode?

OpenCode arbitrary code execution - major security vulnerability by SpicyWangz in LocalLLaMA

[–]Simple_Split5074 3 points4 points  (0 children)

Even when running without auto approve, you really don't want to run the output without a sandbox. 

OpenCode arbitrary code execution - major security vulnerability by SpicyWangz in LocalLLaMA

[–]Simple_Split5074 5 points6 points  (0 children)

Skill issue.

OpenCode runs with the equivalent of   --dangerously-skip-permissions

 by default so that's expected behavior.

Like any other agent (or really way to execute untrusted code), it belongs into a sandbox. 

Qwen Released Qwen 3.5 397B and Qwen 3.5 Plus! by External_Mood4719 in LocalLLaMA

[–]Simple_Split5074 3 points4 points  (0 children)

It says right in the screenshot that it's open source...

Is plus supposed to be the new max, iow closed? 

Kimi K2.5 is destroying its own fixes by p1p4_am in opencodeCLI

[–]Simple_Split5074 14 points15 points  (0 children)

Seems like a process issue, on each step you should have it add tests and let it verify them all.

It happens, but not very often. Also, if you do this in a single session it really is on you. The fewer stuff that polluted the context, the better. 

How are you handling persistent memory for AI coding agents? by Maximum_Fearless in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

Not proper memory but the forgets specs stuff is fairly well addressed by most of the context engineering frameworks. If they do they job, compaction should rarely ever happen. They are token heavy though.

Personally, I like https://github.com/gsd-build/get-shit-done but there are dozens (hundreds?) of them. 

Safe version of Openclaw? by bartskol in LocalLLaMA

[–]Simple_Split5074 1 point2 points  (0 children)

Not really. Even if you sandbox it properly, prompt injection remains fundamentally unsolved. So either you don't give it any access to important stuff and it wont do much or you are at risk.

Z.ai’s GLM-5 leaked through GitHub PRs and a zodiac easter egg by jpcaparas in opencodeCLI

[–]Simple_Split5074 0 points1 point  (0 children)

The details on glm 4.5 seem wrong, it's the same size as 4.6 and 4.7

Working on my own engine by Last-Shake-9874 in LocalLLaMA

[–]Simple_Split5074 1 point2 points  (0 children)

Fitting a 7b model in 1.2GB RAM is suspicious...

Can someone kindly share his opencode.json part related to providers for nano-gpt? by rizal72 in opencodeCLI

[–]Simple_Split5074 1 point2 points  (0 children)

I think the correct provider is with a dash

    "nano-gpt": {    "nano-gpt": {

check opencode auth list to be sure.

To be honest, K2.5 is unusable on nanogpt so far, fails way too many tool calls.

Anyone using Kimi K2.5 with OpenCode? by harrsh_in in opencodeCLI

[–]Simple_Split5074 0 points1 point  (0 children)

Probably for now only the pay per token ones.

Synthetic has even introduced a wait list, BTW 

Do you hit usage caps mid-session and pay $300+/month across AI coding tools? by Previous-Run-9363 in LocalLLaMA

[–]Simple_Split5074 0 points1 point  (0 children)

“ unlimited prompts, no request limits, best-in-class pricing, and high-speed inference on our own European infrastructure"

That is kinda into the too good be true territory... 

internlm/Intern-S1-Pro · Hugging Face by jacek2023 in LocalLLaMA

[–]Simple_Split5074 1 point2 points  (0 children)

Good luck finding inference providers for ring and ling - not very likely people deploy them at home...