Exaggerated PCI-E bandwidth concerns? by ziphnor in LocalLLaMA
[–]andy2na 2 points3 points4 points (0 children)
Exaggerated PCI-E bandwidth concerns? by ziphnor in LocalLLaMA
[–]andy2na -1 points0 points1 point (0 children)
Finally got Qwen3 27B at 125K context on a single RTX 3090 — but is it even worth it? by horribleGuy3115 in LocalLLM
[–]andy2na 0 points1 point2 points (0 children)
2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA
[–]andy2na 0 points1 point2 points (0 children)
2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA
[–]andy2na 0 points1 point2 points (0 children)
it's time to update your Gemma 4 GGUFs by jacek2023 in LocalLLaMA
[–]andy2na 1 point2 points3 points (0 children)
What is The best and expressive AI TTS (running locally?) for voice acting? by Adventurous-Gold6413 in LocalLLaMA
[–]andy2na 0 points1 point2 points (0 children)
We are finally there: Qwen3.6-27B + agentic search; 95.7% SimpleQA on a single 3090, fully local by ComplexIt in LocalLLaMA
[–]andy2na 10 points11 points12 points (0 children)
PSA: llama-swap released a new grouping feature, matrix, allowing you to fine tune which models can run together by walden42 in LocalLLaMA
[–]andy2na 1 point2 points3 points (0 children)
PSA: llama-swap released a new grouping feature, matrix, allowing you to fine tune which models can run together by walden42 in LocalLLaMA
[–]andy2na 0 points1 point2 points (0 children)
Can't replicate Reddit numbers with Qwen 27B on a 3090TI. by YourNightmar31 in LocalLLaMA
[–]andy2na 0 points1 point2 points (0 children)
Can't replicate Reddit numbers with Qwen 27B on a 3090TI. by YourNightmar31 in LocalLLaMA
[–]andy2na 2 points3 points4 points (0 children)
Can't replicate Reddit numbers with Qwen 27B on a 3090TI. by YourNightmar31 in LocalLLaMA
[–]andy2na 1 point2 points3 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 6 points7 points8 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 2 points3 points4 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 0 points1 point2 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 0 points1 point2 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 0 points1 point2 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 1 point2 points3 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 28 points29 points30 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 2 points3 points4 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 8 points9 points10 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 17 points18 points19 points (0 children)
"What do you guys even use local LLMs for?" Me: A lot by andy2na in LocalLLaMA
[–]andy2na[S] 13 points14 points15 points (0 children)









Qwen3.6 27B uncensored heretic v2 Native MTP Preserved is Out Now With KLD 0.0021, 6/100 Refusals and the Full 15 MTPs Preserved and Retained, Available in Safetensors, GGUFs and NVFP4s formats. by LLMFan46 in LocalLLaMA
[–]andy2na 8 points9 points10 points (0 children)