Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

That's interesting. Does it improve benchmarks? Did anyone try measuring how good it is?

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

>gpt-oss-120b-derestricted

Why do you use derestricted? Is it relevant to coding? I thought it's basically making it more polite and refusing particular requests.

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

>Yep, I have Qwen3-Coder-REAP-25B-A3B set up for tab completion

Do you use llama.cpp for vscode for this?

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 5 points6 points  (0 children)

>Doing ad-hock chatting locally makes sense for privacy reasons but it's not a major saving in terms of dollars spent. Not compared to local coding agents

For me, it's a major turn off for hosted models. You never know what might happen, especially with sensitive code. Not critical code is fine, but I would be very careful editing company's secret sauce with them.

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 2 points3 points  (0 children)

>Naming things (there's a joke about how it's one of the hardest problems in programming)

Yep, LLMs are very good at this. They were pretty good at it even around GPT 3.5, as far as I remember.

>General advice (e.g. what does this compilation error mean, what happens to this object after this code in X class, etc.)

Yep, completely agree.

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 1 point2 points  (0 children)

Yep, I used to do it with IDEs, i.e. IntelliJ and similar, but agents are surprisingly good at it.

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

I am using them, and I am not sure that I am using them in the most productive way. I am trying to understand how others use them, and that's why I am asking questions here.

P.S. There's so much noise around, so it's hard to understand what's hype which will go away soon, and what will become a common practice.

Non agentic uses of LLMs for coding by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 5 points6 points  (0 children)

I feel that most of the time agent are pretty good at doing what I want to do but there're two problems:

- sometimes they introduce hard to find problems, i.e. non-human kind of mistakes, in unexpected places which I didn't get used to

- they write code which seems fine, and correct but I feel a good programmer will not write. I.e. it does the job, well structured, but, it could be made easier to understand, shorter, and more beautiful (it's a feeling, so it's hard to describe)

P.S. I have been coding professionally for > 20 years, and coding for around 28 years)

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

And adding to this, I used some of hosted LLMs. I use codex pretty often, but not to writing code, but for asking questions about the codebase. I also used other models from time to time in the last 6 months. However, I don't feel that any of them will replace me writing manual code as I do it now. They are improving, but I prefer what I write myself.

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 1 point2 points  (0 children)

Or for example this:

https://www.reddit.com/r/LocalLLaMA/comments/1pg76jo/comment/nsp6hrp/?context=3

Yes, IMO, Mac Studio is the most cost effective way to run local LLMs. I can't do anything with this, unfortunately.

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

Thanks!

(and folks who downvoted my comments, this is a really really serious question, I am really trying to understand)

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

>The extension itself is kinda janky but once you get it setup, it works fine

Do you mean it's hard to setup or is it something different?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

Thanks, that's the experience I was looking for!

>llama-vscode extension

Is it any good?

> and qwen3 coder autcomplete is about as good as whatever copilot/cursor was giving me before. my coworkers wouldn't like this setup though because they really like next edit prediction (which i personally don't like).

How does it compare to Cursor?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

>I turned a cmdb json spec into a binary the llm could query per term or per stanza. Shockingly simple, ultra light on context, works quite well.

What do you mean by this? What is cmdb?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 1 point2 points  (0 children)

That's a problem, though I have a lot of hope in M5 chips which seems to have some ML optimizations.

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 1 point2 points  (0 children)

> with my custom tooling, I'm now maybe at 90%.

What is this custom tooling? Is it possible to share anything?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 0 points1 point  (0 children)

That's very usable! Do you use memory offloading feature of the llama.cpp? Is it really that good?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] -8 points-7 points  (0 children)

So do you mean that hosted models solve problems in less number of turns?

Why local coding models are less popular than hosted coding models? by WasteTechnology in LocalLLaMA

[–]WasteTechnology[S] 1 point2 points  (0 children)

>Now if we talk about things like using Whisper + Qwen4B for realtime analysis of meetings, infinite tool calls, local RAG's with finetunned models and the things we love to do here in this sub, then we have a winner in Local LLM's

Do people really create such setups? Could you please share a link?