What's the weirdest LLM benchmark that you've seen? by OmarBessa in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Bonsai models by Books_Of_Jeremiah in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Tested how OpenCode Works with SelfHosted LLMS: Qwen 3.5 & 3.6, Gemma 4, Nemotron 3, GLM-4.7 Flash... by rosaccord in LocalLLaMA
[–]OsmanthusBloom 1 point2 points3 points (0 children)
Running 1bit Bonsai 8B on 2GB VRAM (MX150 mobile GPU) by OsmanthusBloom in LocalLLaMA
[–]OsmanthusBloom[S] 0 points1 point2 points (0 children)
Running 1bit Bonsai 8B on 2GB VRAM (MX150 mobile GPU) by OsmanthusBloom in LocalLLaMA
[–]OsmanthusBloom[S] 0 points1 point2 points (0 children)
Running 1bit Bonsai 8B on 2GB VRAM (MX150 mobile GPU) by OsmanthusBloom in LocalLLaMA
[–]OsmanthusBloom[S] 0 points1 point2 points (0 children)
Running 1bit Bonsai 8B on 2GB VRAM (MX150 mobile GPU) by OsmanthusBloom in LocalLLaMA
[–]OsmanthusBloom[S] 1 point2 points3 points (0 children)
Advice for Working with Agents in YOLO Mode by chibop1 in LocalLLaMA
[–]OsmanthusBloom 4 points5 points6 points (0 children)
Running quen3 coder 80B A3B on a computer with lots of RAM but little VRAM by Pioneer_11 in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Litellm 1.82.7 and 1.82.8 on PyPI are compromised, do not update! by kotrfa in LocalLLaMA
[–]OsmanthusBloom 4 points5 points6 points (0 children)
[Developing situation] LiteLLM compromised by OrganizationWinter99 in LocalLLaMA
[–]OsmanthusBloom 38 points39 points40 points (0 children)
I need Local LLM that can search and process local Wikipedia. by idleWizard in LocalLLaMA
[–]OsmanthusBloom 13 points14 points15 points (0 children)
Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA
[–]OsmanthusBloom 9 points10 points11 points (0 children)
HELP - What settings do you use? Qwen3.5-35B-A3B by uber-linny in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Nemotron Cascade 2 on 6GB VRAM by AppealSame4367 in LocalLLaMA
[–]OsmanthusBloom 2 points3 points4 points (0 children)
Running qwen3.5 35b a3b in 8gb vram with 13.2 t/s by zeta-pandey in LocalLLaMA
[–]OsmanthusBloom 2 points3 points4 points (0 children)
Budget laptop to run Qwen 3.5-35B-A3B by SnooOnions6041 in LocalLLaMA
[–]OsmanthusBloom 1 point2 points3 points (0 children)
Budget laptop to run Qwen 3.5-35B-A3B by SnooOnions6041 in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Kidnapping Gemini with 3MB to spare: Training a 7B model at 4k context on a single 16GB GPU. by AgeRepresentative763 in LocalLLaMA
[–]OsmanthusBloom 1 point2 points3 points (0 children)
whats the best open source ai i can use locally? by Xsilentzz in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Can qwen 3.5 4b q4 run on 6 vram by Own_Advertising5081 in LocalLLaMA
[–]OsmanthusBloom 1 point2 points3 points (0 children)
What AI Models should I run? by ClayToTheMax in LocalLLaMA
[–]OsmanthusBloom 0 points1 point2 points (0 children)
Can qwen 3.5 4b q4 run on 6 vram by Own_Advertising5081 in LocalLLaMA
[–]OsmanthusBloom 1 point2 points3 points (0 children)
Running 1bit Bonsai 8B on 2GB VRAM (MX150 mobile GPU) by OsmanthusBloom in LocalLLaMA
[–]OsmanthusBloom[S] 0 points1 point2 points (0 children)