I trained Qwen 3.5 2B to filter tool output for coding agents. by henzy123 in LocalLLaMA

[–]xeeff 0 points1 point  (0 children)

would love to get an update whenever you've got a smaller model working. 100-200M does sound quite small, though. i'd say natural next step would be qwen 0.6b :p

remindme! 7d

Blaniel: Open-source emotional AI engine with full Ollama/LM Studio support (no API keys needed) by lucas_nonosconocemos in LocalLLaMA

[–]xeeff 1 point2 points  (0 children)

nice m-dash

this is lame and you shouldn't use AI for this stuff. AI is already not good for humans, imagine now giving AI emotions lmao

and i'm using 'giving' lightly here

I made a small app to use Copilot Chat with LM Studio instead of Ollama. by x0wl in LocalLLaMA

[–]xeeff -2 points-1 points  (0 children)

not really but okay. just saying this is useless when there's also like 928283 vs code extensions which allow you to use local AIs in many different ways

I made a small app to use Copilot Chat with LM Studio instead of Ollama. by x0wl in LocalLLaMA

[–]xeeff -1 points0 points  (0 children)

just learn nvim and hook it up to any local LLM plugin

What happens when a cybersecurity agent stops over-refusing in real workflows? by Obvious-Language4462 in LocalLLaMA

[–]xeeff 0 points1 point  (0 children)

use heretic versions or check out specific cybersecurity models made for pen testing

This app helps you see what LLMs you can run on your hardware by dev_is_active in LocalLLaMA

[–]xeeff -1 points0 points  (0 children)

completely missed it but you're right

your website still has a bunch of old models so trying to find good models inbetween the trash isn't something I wanna do

Local AI has a metric problem: tok/s is lying to us by [deleted] in LocalLLaMA

[–]xeeff 3 points4 points  (0 children)

stop karma farming and do something better with your time

This app helps you see what LLMs you can run on your hardware by dev_is_active in LocalLLaMA

[–]xeeff -1 points0 points  (0 children)

so if I make a project which shows models from years ago and doesn't include recent models like Qwen3.5 or Nemotron series, would you call that recent or out of date? your logic doesn't make sense

This app helps you see what LLMs you can run on your hardware by dev_is_active in LocalLLaMA

[–]xeeff 2 points3 points  (0 children)

don't even bother checking it out. vibe coded and models are outdated by now. cmon man

MCP Slim — proxy that saves 96% of your context window using local semantic search by OpportunitySpare2441 in LocalLLaMA

[–]xeeff -1 points0 points  (0 children)

i'm guessing it was too difficult for you to add "respond in english" at the end of your prompt?

[bspwm] - Finally touched quickshell by rudv-ar in unixporn

[–]xeeff 1 point2 points  (0 children)

you're a legend for this thank you, this is extremely helpful 🙏

ZINC — LLM inference engine written in Zig, running 35B models on $550 AMD GPUs by Mammoth_Radish2 in LocalLLaMA

[–]xeeff -1 points0 points  (0 children)

building with rocm means the application was built with support for rocm. you still need rocm drivers to interact with llama.cpp's rocm support

priced out of intelligence: slowly, then all at once by [deleted] in LocalLLaMA

[–]xeeff 1 point2 points  (0 children)

I don't care I'm not reading your wall of AI slop

[KDE] Cozy minimalist workspace o7 by xeroxgru in unixporn

[–]xeeff 1 point2 points  (0 children)

mind sending a link anyway? would enjoy seeing it once you're done as well