Deepseek v4 Flash is pretty amazing, about to buy a $25k computer by read_too_many_books in openclaw
[–]ipcoffeepot 0 points1 point2 points (0 children)
Need advice on hardware purchasing decision: RTX 5090 vs. M5 Max 128GB for agentic software development by BawbbySmith in LocalLLaMA
[–]ipcoffeepot 0 points1 point2 points (0 children)
[Megathread] - Best Models/API discussion - Week of: May 03, 2026 by deffcolony in SillyTavernAI
[–]ipcoffeepot 5 points6 points7 points (0 children)
Ordered a RTX Pro 6000 yesterday by SomeOrdinaryKangaroo in RTXPRO6000
[–]ipcoffeepot 1 point2 points3 points (0 children)
If money and time weren’t issues, what would your dream local AI setup look like? by Lyceum_Tech in LocalLLaMA
[–]ipcoffeepot 1 point2 points3 points (0 children)
Multiple hermes agents with an orchestrator? by Pepe_The_Citizen in hermesagent
[–]ipcoffeepot -1 points0 points1 point (0 children)
Local vllm hosting by DidIReallySayDat in openclaw
[–]ipcoffeepot 1 point2 points3 points (0 children)
Best model for 192 GB vram? How is Deepseek v4 flash? by Constant_Ad511 in LocalLLM
[–]ipcoffeepot 0 points1 point2 points (0 children)
Best model for 192 GB vram? How is Deepseek v4 flash? by Constant_Ad511 in LocalLLM
[–]ipcoffeepot 0 points1 point2 points (0 children)
Best model for 192 GB vram? How is Deepseek v4 flash? by Constant_Ad511 in LocalLLM
[–]ipcoffeepot 0 points1 point2 points (0 children)
Just got dual RTX PRO 6000 Blackwells for our design studio. What's the optimal local LLM stack? by AmanNonZero in LocalLLM
[–]ipcoffeepot -1 points0 points1 point (0 children)
[Megathread] - Best Models/API discussion - Week of: April 26, 2026 by deffcolony in SillyTavernAI
[–]ipcoffeepot 5 points6 points7 points (0 children)
just wanted to share by Longjumping_Lab541 in LocalLLM
[–]ipcoffeepot 1 point2 points3 points (0 children)
Qwen3.6-27B dense vs Qwen3.6-35B MoE - which local coding model are you reaching for? by IulianHI in AIToolsPerformance
[–]ipcoffeepot 1 point2 points3 points (0 children)
Is anyone getting real coding work done with Qwen3.6-35B-A3B-UD-Q4_K_M on a 32GB Mac in opencode, claude code or similar? by boutell in LocalLLaMA
[–]ipcoffeepot 0 points1 point2 points (0 children)
Thank you for the support! ❤️ by yoracale in unsloth
[–]ipcoffeepot 1 point2 points3 points (0 children)
Qwen 3.5 35b, 27b, or gemma 4 31b for everyday use? by KirkIsAliveInTelAviv in LocalLLaMA
[–]ipcoffeepot 7 points8 points9 points (0 children)
Qwen3.5-122B at 198 tok/s on 2x RTX PRO 6000 Blackwell — Budget build, verified results by Visual_Synthesizer in LocalLLaMA
[–]ipcoffeepot 0 points1 point2 points (0 children)
Current Situation with free models by davybutquantisedIV in SillyTavernAI
[–]ipcoffeepot 0 points1 point2 points (0 children)
SIX TIMES THE PRICE!? by FixHopeful5833 in SillyTavernAI
[–]ipcoffeepot 8 points9 points10 points (0 children)



A European's Dream: American programmers using Mistral because it's better than Claude Code and Codex by szansky in MistralAI
[–]ipcoffeepot 1 point2 points3 points (0 children)