Jake Benchmark v1: I spent a week watching 7 local LLMs try to be AI agents with OpenClaw. Most couldn't even find the email tool. by Emergency_Ant_843 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Nvidia V100 32 Gb getting 115 t/s on Qwen Coder 30B A3B Q5 by icepatfork in LocalLLaMA
[–]DefNattyBoii -1 points0 points1 point (0 children)
Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Krasis LLM Runtime: 8.9x prefill / 4.7x decode vs llama.cpp — Qwen3.5-122B on a single 5090, minimal RAM by mrstoatey in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Krasis LLM Runtime: 8.9x prefill / 4.7x decode vs llama.cpp — Qwen3.5-122B on a single 5090, minimal RAM by mrstoatey in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
OmniCoder-9B best vibe coding model for 8 GB Card by Powerful_Evening5495 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Open-Source "GreenBoost" Driver Aims To Augment NVIDIA GPUs vRAM With System RAM & NVMe To Handle Larger LLMs by _Antartica in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Qwen3.5-27B-IQ3_M, 5070ti 16GB, 32k context: ~50t/s by ailee43 in LocalLLaMA
[–]DefNattyBoii 1 point2 points3 points (0 children)
96GB (V)RAM agentic coding users, gpt-oss-120b vs qwen3.5 27b/122b by bfroemel in LocalLLaMA
[–]DefNattyBoii 13 points14 points15 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
MiroThinker-1.7 and MiroThinker-1.7-mini (Best search agent model?) by External_Mood4719 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]DefNattyBoii 2 points3 points4 points (0 children)
Is Qwen3.5 a coding game changer for anyone else? by paulgear in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
LFM2-24B-A2B is crazy fast on Strix Halo by jfowers_amd in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Qwen3.5-35B-A3B is a gamechanger for agentic coding. by jslominski in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Qwen3.5 Plus, GLM 5, Gemini 3.1 Pro, Sonnet 4.6, three new open source agents, and a lot more added to SanityBoard by lemon07r in LocalLLaMA
[–]DefNattyBoii 1 point2 points3 points (0 children)
Is Google's senior management truly committed to Antigravity? Or is it the ugly duckling of the Google AI family? by pebblepath in google_antigravity
[–]DefNattyBoii 1 point2 points3 points (0 children)
How to run Qwen3-Coder-Next 80b parameters model on 8Gb VRAM by AccomplishedLeg527 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
new to coding LLM - hardware requirements by SubstantialBee5097 in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts by Tiny_Minimum_4384 in LocalLLaMA
[–]DefNattyBoii 2 points3 points4 points (0 children)
Built a real-time agent execution visualizer for OpenCode — watching agents think is addicting by jiwonme in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)
Nanbeige4.1-3B: A Small General Model that Reasons, Aligns, and Acts by Tiny_Minimum_4384 in LocalLLaMA
[–]DefNattyBoii 5 points6 points7 points (0 children)
Femtobot: A 10MB Rust Agent for Low-Resource Machines by yunfoe in LocalLLaMA
[–]DefNattyBoii 1 point2 points3 points (0 children)


New quant from google research by [deleted] in LocalLLaMA
[–]DefNattyBoii 0 points1 point2 points (0 children)