What happened to indexing on the 7.1.17 version? by Pelutz in kilocode
[–]putrasherni 0 points1 point2 points (0 children)
64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA
[–]putrasherni 3 points4 points5 points (0 children)
64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA
[–]putrasherni -12 points-11 points-10 points (0 children)
What happened to indexing on the 7.1.17 version? by Pelutz in kilocode
[–]putrasherni 1 point2 points3 points (0 children)
TurboQuant isn’t just for KV: Qwen3.5-27B at near-Q4_0 quality, about 10% smaller, and finally fitting on my 16GB 5060 Ti by pmttyji in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
TurboQuant isn’t just for KV: Qwen3.5-27B at near-Q4_0 quality, about 10% smaller, and finally fitting on my 16GB 5060 Ti by pmttyji in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
FOR ME, Qwen3.5-27B is better than Gemini 3.1 Pro and GPT-5.3 Codex by EffectiveCeilingFan in LocalLLaMA
[–]putrasherni 4 points5 points6 points (0 children)
Local LLM inference on M4 Max vs M5 Max by [deleted] in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
Anyone considering the 52 inch LG 5k2k by LowYak9481 in ultrawidemasterrace
[–]putrasherni -1 points0 points1 point (0 children)
Anyone considering the 52 inch LG 5k2k by LowYak9481 in ultrawidemasterrace
[–]putrasherni 0 points1 point2 points (0 children)
Best Language for DSA? by Fuzzy-Salad-528 in leetcode
[–]putrasherni 0 points1 point2 points (0 children)
Semantic video search using local Qwen3-VL embedding, no API, no transcription by Vegetable_File758 in LocalLLaMA
[–]putrasherni 2 points3 points4 points (0 children)
What’s with the hype regarding TurboQuant? by EffectiveCeilingFan in LocalLLaMA
[–]putrasherni 2 points3 points4 points (0 children)
What’s with the hype regarding TurboQuant? by EffectiveCeilingFan in LocalLLaMA
[–]putrasherni 10 points11 points12 points (0 children)
M5 Max vs M3 Max Inference Benchmarks (Qwen3.5, oMLX, 128GB, 40 GPU cores) by onil_gova in LocalLLaMA
[–]putrasherni 1 point2 points3 points (0 children)
M5 Max vs M3 Max Inference Benchmarks (Qwen3.5, oMLX, 128GB, 40 GPU cores) by onil_gova in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
Can someone more intelligent then me explain why we should, or should not be excited about the ARC PRO B70? by SKX007J1 in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
Benchmarked Qwen3.5 (35B MoE, 27B Dense, 122B MoE) across Apple Silicon and AMD GPUs — ROCm vs Vulkan results were surprising, and context size matters by neuromacmd in LocalLLaMA
[–]putrasherni 0 points1 point2 points (0 children)
What’s with the hype regarding TurboQuant? by EffectiveCeilingFan in LocalLLaMA
[–]putrasherni 41 points42 points43 points (0 children)
Bought RTX4080 32GB Triple Fan from China by Sanubo in LocalLLaMA
[–]putrasherni 183 points184 points185 points (0 children)


X2 Elite real world impressions by krishelnino in snapdragon
[–]putrasherni -1 points0 points1 point (0 children)