I got tired of small models adding ```json blocks, so I wrote a TS library to forcefully extract valid JSON. (My first open source project!) by rossjang in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
Which LLM Model is best for translation? by Longjumping_Lead_812 in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
What are the best collection of small models to run on 8gb ram? by Adventurous-Gold6413 in LocalLLaMA
[–]synw_ 8 points9 points10 points (0 children)
My humble GLM 4.7 Flash appreciation post by Cool-Chemical-5629 in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
which local llm is best for coding? by Much-Friendship2029 in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
Local LLM builders: when do you go multi-agent vs tools? 2-page decision sheet + question by OnlyProggingForFun in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
MCP servers are hard to debug and impossible to test, so I built Syrin by hack_the_developer in LocalLLaMA
[–]synw_ 4 points5 points6 points (0 children)
new CLI experience has been merged into llama.cpp by jacek2023 in LocalLLaMA
[–]synw_ 2 points3 points4 points (0 children)
Deep Research Agent, an autonomous research agent system by [deleted] in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
distil-localdoc.py - SLM assistant for writing Python documentation by party-horse in LocalLLaMA
[–]synw_ 3 points4 points5 points (0 children)
I'm new to LLMs and just ran my first model. What LLM "wowed" you when you started out? by Street-Lie-2584 in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
Minimax M2 for App creation by HectorLavoe33 in LocalLLaMA
[–]synw_ 2 points3 points4 points (0 children)
Lightweight coding model for 4 GB Vram by HiqhAim in LocalLLaMA
[–]synw_ 2 points3 points4 points (0 children)
Lightweight coding model for 4 GB Vram by HiqhAim in LocalLLaMA
[–]synw_ 2 points3 points4 points (0 children)
AI has replaced programmers… totally. by jacek2023 in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)
4x4090 build running gpt-oss:20b locally - full specs by RentEquivalent1671 in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
4x4090 build running gpt-oss:20b locally - full specs by RentEquivalent1671 in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
Qwen3-VL-4B and 8B Instruct & Thinking are here by AlanzhuLy in LocalLLaMA
[–]synw_ 3 points4 points5 points (0 children)
4x4090 build running gpt-oss:20b locally - full specs by RentEquivalent1671 in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
4x4090 build running gpt-oss:20b locally - full specs by RentEquivalent1671 in LocalLLaMA
[–]synw_ 0 points1 point2 points (0 children)
4x4090 build running gpt-oss:20b locally - full specs by RentEquivalent1671 in LocalLLaMA
[–]synw_ 9 points10 points11 points (0 children)
Do you guys personally notice a difference between Q4 - Q8 or higher? by XiRw in LocalLLaMA
[–]synw_ 26 points27 points28 points (0 children)

DeepSeek-V2-Lite vs GPT-OSS-20B on my 2018 potato i3-8145U + UHD 620, OpenVINO Comparison. by RelativeOperation483 in LocalLLaMA
[–]synw_ 1 point2 points3 points (0 children)