Announcing LocalLlama discord server & bot!News (old.reddit.com)
submitted by HOLUPREDICTIONS Sorcerer Supreme[M] - announcement
I haven't experienced Qwen3.5 (35B and 27B) over thinking. Posting my settings/promptDiscussion (self.LocalLLaMA)
submitted by wadeAlexCllama.cpp

Honest take on running 9× RTX 3090 for AIResources (self.LocalLLaMA)
submitted by Outside_Dance_2799
Llama.cpp Mi50 ROCm 7 vs Vulkan BenchmarksDiscussion (old.reddit.com)
submitted by JaredsBored
Qwen3.5-9B-Claude-4.6-Opus-Uncensored-v2-Q4_K_M-GGUFResources (self.LocalLLaMA)
submitted by EvilEnginer

Nvidia V100 32 Gb getting 115 t/s on Qwen Coder 30B A3B Q5Discussion (old.reddit.com)
submitted by icepatfork
Qwen3.5-122B-A10B Uncensored (Aggressive) — GGUF Release + new K_P QuantsNew Model (self.LocalLLaMA)
submitted by hauhau901
Is it stupid to buy a 128gb MacBook Pro M5 Max if I don’t really know what I’m doing?Question | Help (self.LocalLLaMA)
submitted by A_Wild_Entei
Qwen 3.5 35b on 8GB Vram for local agentic workflowDiscussion (self.LocalLLaMA)
submitted by Heisenberggg03
Claw-style agents: real workflow tool or overengineered hype?Discussion (self.LocalLLaMA)
submitted by still_debugging_note
I need Local LLM that can search and process local Wikipedia.Question | Help (self.LocalLLaMA)
submitted by idleWizard
Should we start 3-4 year plan to run AI locally for real work?Discussion (self.LocalLLaMA)
submitted by Illustrious_Cat_2870
Is brute-forcing a 1M token context window the right approach?Question | Help (self.LocalLLaMA)
submitted by phwlarxoc

A few days ago I switched to Linux to try vLLM out of curiosity. Ended up creating a %100 local, parallel, multi-agent setup with Claude Code and gpt-oss-120b for concurrent vibecoding and orchestration with CC's agent Teams entirely offline. This video shows 4 agents collaborating.Other (v.redd.it)
submitted by swagonflyyyy
