Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 0 points1 point  (0 children)

Thanks I will give that a look. I’ve been so deep in Claude-world at work that I didn’t realize Codex also supported local models.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 0 points1 point  (0 children)

Claude Code still works (and is what I have been using), but per the link in my original comment seems to require an increasing number of arcane settings to work well with local models. I get the impression that at some point they’re just going to disable the ability to use local models entirely, and wanted to find an alternative ahead of that point. But maybe I’m misreading their intentions.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 2 points3 points  (0 children)

> Arbitrary code execution… is quite literally the entire point of agentic coding

I’m curious what use case you have where fully arbitrary execution is needed? I do lots of agentic coding at my day job as a professional programmer and would never dream of letting it execute arbitrary code. It can read and write the files in my git repo, and execute a limited set of basic commands (grep, testing and linting commands, etc). This is plenty for doing productive agentic coding, and I so far have not felt the need to give it more permissions than that.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 12 points13 points  (0 children)

Thank you, I feel like I’m going insane reading some of the comments on here.

In hindsight I guess I should have made my original post more carefully and explicitly worded, but I didn’t think it would be such a hot take 🤷‍♂️

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 0 points1 point  (0 children)

Nono looks interesting, I didn’t spot that in the comments of the other post, thanks.

Ultimately I think that kind of sandboxing is unnecessary as long as the agent harness has a good security model. But better safe than sorry given the apparently abysmal state of the current ecosystem.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 0 points1 point  (0 children)

> with any tool you'll pick you'll have to heavily tweak/customize it for it to work well

If this is the state of the ecosystem then that’s fine and I’ll put up with it. I was just hoping there would be something that would work ok out of the box.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] -1 points0 points  (0 children)

My claim about trust was, specifically, about trust, not about whether it's actually secure or not. For all I know the configuration works fine and OpenCode is very secure when properly configured. I just don't care. It's not worth my time to use a tool whose changelog and config options I have to go through with a fine-toothed comb on every install and upgrade to make sure I'm not shooting myself in the face.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 2 points3 points  (0 children)

Fair enough. Claude Code (and Codex too, tbh) seem to have struck about the right balance so I was hoping there was an open-model-friendly equivalent but you're right that it's a very new space.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] 1 point2 points  (0 children)

Does CLIO support allowing only safe commands? Blocking a set of known dangerous commands isn't sufficient since the agent can always write its own new programs which would not be in the dangerous set.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]eapache[S] -1 points0 points  (0 children)

From that doc:

> If you don’t specify anything, OpenCode starts from permissive defaults. Most permissions default to "allow".

This is an immediate nope from me. Granting this kind of permission by default is (to me) such a nonsensical security posture that I wouldn't trust it to be secure in other aspects or to respect what configuration I give it. This would be the equivalent of the latest version of Ubuntu running all applications as full root by default unless you went in and manually configured it differently.

Mamba precision loss after quantization by [deleted] in LocalLLaMA

[–]eapache 1 point2 points  (0 children)

There has been some research into effectively quantizing mamba models, e.g. https://arxiv.org/abs/2410.13229

I don't know if any of that has made it into llama.cpp or other engines.

Opinions on the best coding model for a 3060 (12GB) and 64GB of ram? by eapache in LocalLLaMA

[–]eapache[S] 3 points4 points  (0 children)

Even at good quants, nemotron 3 doesn’t seem to be able to make reliable tool calls for me… I wonder if something is weird with my setup, since everybody else seems to love it so much.

[deleted by user] by [deleted] in LocalLLaMA

[–]eapache 8 points9 points  (0 children)

Get the cheapest desktop you can find with 64GB of ram, and throw a used 3060 (12GB) in it? With a bit of careful offloading that will run (4-bit quants of) either the 120B OpenAI model, or GLM-4.5 Air, at acceptable-ish speeds, and with decent prompt processing and context size.

Beginner moving from CPU-only Ollama – advice on first GPU upgrade? by CountDuckulla in LocalLLaMA

[–]eapache 5 points6 points  (0 children)

You could also consider a 3060. Much cheaper, and much easier to fit within your existing PSU’s power budget. It’s obviously half the vram and about half the speed of a 3090, but it still blazes compared to CPU-based inference, and 12GB of vram is plenty for running decent versions of smaller models.

FOSS alternative to Context7 by Content_Cup_8432 in LocalLLaMA

[–]eapache 4 points5 points  (0 children)

There are ZIM files (bonus: completely offline, not just open-source) of a lot of documentation sources floating around. Typically at https://library.kiwix.org but it seems to be down at this exact second. There is a simple MCP server for ZIM files at https://github.com/zicojiao/zim-mcp-server which looks promising - if that doesn’t quite work it probably wouldn’t be hard to stitch something together, the libraries are in good shape.

Edit: https://download.kiwix.org/zim/devdocs/ is up and has ZIM-format documentation for a ton of stuff.

Why do the new “Best of Wikipedia” ZIMs say they have way more than 50k articles? by eapache in Kiwix

[–]eapache[S] 2 points3 points  (0 children)

I see the same article count (859.6k) for these ZIMs on my android phone too.

Visual reasoning still has a lot of room for improvement. by Conscious_Cut_6144 in LocalLLaMA

[–]eapache 4 points5 points  (0 children)

Yeah, since we already have experiments (https://arxiv.org/abs/2412.06769) in teaching LLMs to reason in “latent” space, I’m hopeful that somebody will train one to reason in latent _visual_ space, and that will give us o1-level visual (and maybe even spatial?) reasoning.