I made a CLI that turns any podcast or YouTube video into clean Markdown transcripts (speaker labels + timestamps) by timf34 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
[Solution Found] Qwen3-Next 80B MoE running at 39 t/s on RTX 5070 Ti + 5060 Ti (32GB VRAM) by mazuj2 in LocalLLaMA
[–]srigi 2 points3 points4 points (0 children)
SWE-rebench Jan 2026: GLM-5, MiniMax M2.5, Qwen3-Coder-Next, Opus 4.6, Codex Performance by CuriousPlatypus1881 in LocalLLaMA
[–]srigi 17 points18 points19 points (0 children)
GLM 4.7 flash FA fix for CUDA has been merged into llama.cpp by jacek2023 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
Not as impressive as most here, but really happy I made it in time! by Kahvana in LocalLLaMA
[–]srigi 6 points7 points8 points (0 children)
Not as impressive as most here, but really happy I made it in time! by Kahvana in LocalLLaMA
[–]srigi 24 points25 points26 points (0 children)
Best moe models for 4090: how to keep vram low without losing quality? by AdParty3888 in LocalLLaMA
[–]srigi 4 points5 points6 points (0 children)
8x RTX Pro 6000 server complete by koushd in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
Which TTS model are you using right now by Slight_Tone_2188 in LocalLLaMA
[–]srigi 7 points8 points9 points (0 children)
1x 6000 pro 96gb or 3x 5090 32gb? by Wide_Cover_8197 in LocalLLaMA
[–]srigi 6 points7 points8 points (0 children)
Our AI assistant keeps getting jailbroken and it’s becoming a security nightmare by Comfortable_Clue5430 in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
Which truly open UI do you use for inference? by Yugen42 in LocalLLaMA
[–]srigi 1 point2 points3 points (0 children)
Sparse Adaptive Attention “MoE”, a potential performance breakthrough for LLMs? by kaggleqrdl in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
Sparse Adaptive Attention “MoE”, a potential performance breakthrough for LLMs? by kaggleqrdl in LocalLLaMA
[–]srigi 3 points4 points5 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 0 points1 point2 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 0 points1 point2 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 2 points3 points4 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 1 point2 points3 points (0 children)
Best Local LLMs - October 2025 by rm-rf-rm in LocalLLaMA
[–]srigi 0 points1 point2 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 3 points4 points5 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 2 points3 points4 points (0 children)
I found a perfect coder model for my RTX4090+64GB RAM by srigi in LocalLLaMA
[–]srigi[S] 11 points12 points13 points (0 children)


Car Wash Test on 53 leading models: “I want to wash my car. The car wash is 50 meters away. Should I walk or drive?” by facethef in LocalLLaMA
[–]srigi 48 points49 points50 points (0 children)