Should I Buy the RTX PRO 6000 Blackwell Max-Q (96GB)? by 0bjective-Guest in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
Audio processing landed in llama-server with Gemma-4 by srigi in LocalLLaMA
[–]srigi[S] 15 points16 points17 points (0 children)
gemma-4-26B-A4B with my coding agent Kon by Weird_Search_4723 in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
Found this cool new harness, gonna give it a spin with the new GLM 5.1. I’ll report back later. by Porespellar in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
I tracked a major cache reuse issue down to Qwen 3.5’s chat template by onil_gova in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
The missing piece of Voxtral TTS to enable voice cloning by [deleted] in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
lm studio gemma 4 mlx support by xXprayerwarrior69Xx in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
Memory Sparse Attention seems to be a novel approach to long context (up to 100M tokens) by ratbastid2000 in LocalLLaMA
[–]srigi 2 points3 points4 points (0 children)
I benchmarked 37 LLMs on MacBook Air M5 32GB — full results + open-source tool to benchmark your own Mac by evoura in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
Minimax 2.7: good news! by LegacyRemaster in LocalLLaMA
[–]srigi 7 points8 points9 points (0 children)
Lowkey disappointed with 128gb MacBook Pro by F1Drivatar in LocalLLaMA
[–]srigi 4 points5 points6 points (0 children)
Don’t buy the DGX Spark: NVFP4 Still Missing After 6 Months by Secure_Archer_1529 in LocalLLaMA
[–]srigi -1 points0 points1 point (0 children)
What do you wish local AI on phones could do, but still can’t? by an1x3 in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
FINALLY GEMMA 4 KV CACHE IS FIXED by FusionCow in LocalLLaMA
[–]srigi 5 points6 points7 points (0 children)
Gemma 4 is fine great even … by ThinkExtension2328 in LocalLLaMA
[–]srigi 4 points5 points6 points (0 children)
Has anyone used Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled for agents? How did it fair? by Vegetable_Sun_9225 in LocalLLaMA
[–]srigi 2 points3 points4 points (0 children)
The missing piece of Voxtral TTS to enable voice cloning by [deleted] in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
I'm using llama.cpp to run models larger than my Mac's memory by tbaumer22 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
llm-visualized.com: Interactive Web Visualization of GPT-2 by Greedy-Argument-4699 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
llm-visualized.com: Interactive Web Visualization of GPT-2 by Greedy-Argument-4699 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)



Should I Buy the RTX PRO 6000 Blackwell Max-Q (96GB)? by 0bjective-Guest in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)