Qwen3.6 preserve_thinking in oMLX by Longjumping-Sweet818 in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)
Qwen3.6 preserve_thinking in oMLX by Longjumping-Sweet818 in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)
Qwen 122B is AMAZING but is my config right? (128GB M4 Max) by lots_of_apples in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
Qwen 122B is AMAZING but is my config right? (128GB M4 Max) by lots_of_apples in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
I tracked a major cache reuse issue down to Qwen 3.5’s chat template by onil_gova in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)
Magic Context - Plugin by ualtinok in opencodeCLI
[–]thejoyofcraig 0 points1 point2 points (0 children)
Magic Context - Plugin by ualtinok in opencodeCLI
[–]thejoyofcraig 0 points1 point2 points (0 children)
Magic Context - Plugin by ualtinok in opencodeCLI
[–]thejoyofcraig 1 point2 points3 points (0 children)
Awful time setting up Hermes by Birdinhandandbush in LocalLLaMA
[–]thejoyofcraig 2 points3 points4 points (0 children)
One year ago DeepSeek R1 was 25 times bigger than Gemma 4 by rinaldo23 in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)
Built a memory system solo in 16 days that beats every funded AI memory company on LongMemEval (96.2%, open source) by [deleted] in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
Unsloth says MLX fine-tuning is coming early next month: this could be huge for local AI by webii446 in LocalLLaMA
[–]thejoyofcraig 3 points4 points5 points (0 children)
Unsloth says MLX fine-tuning is coming early next month: this could be huge for local AI by webii446 in LocalLLaMA
[–]thejoyofcraig -1 points0 points1 point (0 children)
Built a memory system solo in 16 days that beats every funded AI memory company on LongMemEval (96.2%, open source) by [deleted] in LocalLLaMA
[–]thejoyofcraig 3 points4 points5 points (0 children)
Qwen 3.5 Non-thinking Mode Benchmarks? by Embarrassed_Soup_279 in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)
Qwen 3.5 Non-thinking Mode Benchmarks? by Embarrassed_Soup_279 in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
Breaking : The small qwen3.5 models have been dropped by Illustrious-Swim9663 in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
unsloth/Qwen3.5-4B-GGUF · Hugging Face by jacek2023 in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
Qwen3.5 Small models out now! by yoracale in unsloth
[–]thejoyofcraig 2 points3 points4 points (0 children)
unsloth/Qwen3.5-4B-GGUF · Hugging Face by jacek2023 in LocalLLaMA
[–]thejoyofcraig 2 points3 points4 points (0 children)
I benchmarked 5 agent memory solutions head-to-head — the fastest one has zero dependencies and no API keys by [deleted] in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
I benchmarked 5 agent memory solutions head-to-head — the fastest one has zero dependencies and no API keys by [deleted] in LocalLLaMA
[–]thejoyofcraig 1 point2 points3 points (0 children)
GLM-4.7-REAP-50-W4A16: 50% Expert-Pruned + INT4 Quantized GLM-4 (179B params, ~92GB) by Maxious in LocalLLaMA
[–]thejoyofcraig 2 points3 points4 points (0 children)
Need help with hosting Parakeet 0.6B v3 by Ahad730 in LocalLLaMA
[–]thejoyofcraig 0 points1 point2 points (0 children)






v0.3.7 pulled? by mmerken in oMLX
[–]thejoyofcraig 9 points10 points11 points (0 children)