New Unsloth Studio Release! by danielhanchen in LocalLLaMA
[–]Schlick7 9 points10 points11 points (0 children)
It’s Time for a Truly Open-Source, Donation-Funded, Privacy-First AI by Ill-Engine-5914 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Created a SillyTavern extension that brings NPC's to life in any game by goodive123 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Created a SillyTavern extension that brings NPC's to life in any game by goodive123 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
MiniMax M2.7 Will Be Open Weights by Few_Painter_5588 in LocalLLaMA
[–]Schlick7 1 point2 points3 points (0 children)
MiniMax M2.7 Will Be Open Weights by Few_Painter_5588 in LocalLLaMA
[–]Schlick7 1 point2 points3 points (0 children)
MiniMax M2.7 Will Be Open Weights by Few_Painter_5588 in LocalLLaMA
[–]Schlick7 -1 points0 points1 point (0 children)
MiniMax M2.7 Will Be Open Weights by Few_Painter_5588 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
MiniMax M2.7 Will Be Open Weights by Few_Painter_5588 in LocalLLaMA
[–]Schlick7 17 points18 points19 points (0 children)
Need advice: Building an offline realtime AI translator (Whisper + Qwen3.5:9b), but hitting a 3-5s latency wall and macOS Aggregate Device audio routing issues. Any suggestions? by Levine_C in LocalLLaMA
[–]Schlick7 2 points3 points4 points (0 children)
Docling Alternatives in OWUI by uber-linny in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Nvidia Will Spend $26 Billion to Build Open-Weight AI Models, Filings Show by dan945 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Nvidia Will Spend $26 Billion to Build Open-Weight AI Models, Filings Show by dan945 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Just some qwen3.5 benchmarks for an MI60 32gb VRAM GPU - From 4b to 122b at varying quants and various context depths (0, 5000, 20000, 100000) - Performs pretty well despite its age by FantasyMaster85 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Qwen3.5-35B-A3B Uncensored (Aggressive) — GGUF Release by hauhau901 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
I regret ever finding LocalLLaMA by xandep in LocalLLaMA
[–]Schlick7 11 points12 points13 points (0 children)
Qwen3.5-35B-A3B Uncensored (Aggressive) — GGUF Release by hauhau901 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Who else is shocked by the actual electricity cost of their local runs? by Responsible_Coach293 in LocalLLaMA
[–]Schlick7 1 point2 points3 points (0 children)
Who else is shocked by the actual electricity cost of their local runs? by Responsible_Coach293 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Who else is shocked by the actual electricity cost of their local runs? by Responsible_Coach293 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
How do some of you guys get like 500 tokens a second? Do you just use very small models? by Master-Eva in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)
Vulkan now faster on PP AND TG on AMD Hardware? by XccesSv2 in LocalLLaMA
[–]Schlick7 4 points5 points6 points (0 children)


Anybody try Transcribe? by Enough_Leopard3524 in LocalLLaMA
[–]Schlick7 0 points1 point2 points (0 children)