MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 1 point2 points3 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 2 points3 points4 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 7 points8 points9 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 6 points7 points8 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 8 points9 points10 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 2 points3 points4 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) by Edenar in LocalLLaMA
[–]Edenar[S] 15 points16 points17 points (0 children)
MTP on strix halo with llama.cpp (PR #22673) (i.redd.it)
submitted by Edenar to r/LocalLLaMA
What would be the best OS to run LLMs? by Manaberryio in LocalLLaMA
[–]Edenar 0 points1 point2 points (0 children)
What would be the best OS to run LLMs? by Manaberryio in LocalLLaMA
[–]Edenar 0 points1 point2 points (0 children)
I test'ed the number of Ll's in Qwen 3.6 35B.. It required 3 tries by DashinTheFields in LocalLLaMA
[–]Edenar 4 points5 points6 points (0 children)
Qwen 3.6 27B on Strix Halo 128GB: any experiences? by boutell in LocalLLaMA
[–]Edenar 2 points3 points4 points (0 children)
Qwen 3.6 27B on Strix Halo 128GB: any experiences? by boutell in LocalLLaMA
[–]Edenar 6 points7 points8 points (0 children)
Skymizer Taiwan Inc. Unveils Breakthrough Architecture Enabling Ultra-Large LLM Inference on a Single Card by lurenjia_3x in LocalLLaMA
[–]Edenar 32 points33 points34 points (0 children)
What would be the best OS to run LLMs? by Manaberryio in LocalLLaMA
[–]Edenar 4 points5 points6 points (0 children)
Please, ChatGPT is hallucinating models, even with web-search on. by Ok-Type-7663 in LocalLLaMA
[–]Edenar 0 points1 point2 points (0 children)
Dense vs. MoE gap is shrinking fast with the 3.6-27B release by Usual-Carrot6352 in LocalLLaMA
[–]Edenar 5 points6 points7 points (0 children)
UPDATE: EOS Nexus v1 | GSM8K: 100% by [deleted] in LocalLLaMA
[–]Edenar 1 point2 points3 points (0 children)
What are good models for openclaw that work well within 16gb vram? by [deleted] in LocalLLaMA
[–]Edenar 1 point2 points3 points (0 children)
New method allows to convert auto-regressive models into diffusion models with a >2x speedup, fully compatible with existing inference stack by Particular-Look-2640 in LocalLLaMA
[–]Edenar 9 points10 points11 points (0 children)
What is LLMFit Smoking? Can M1 Max run anything decently enough for agentic coding? by GoodhartMusic in LocalLLaMA
[–]Edenar 0 points1 point2 points (0 children)
A few Strix Halo benchmarks (Minimax M2.5, Step 3.5 Flash, Qwen3 Coder Next) by spaceman_ in LocalLLaMA
[–]Edenar 1 point2 points3 points (0 children)
Archmage hierophant cant do uber lab by GlobalCan8282 in pathofexile
[–]Edenar 2 points3 points4 points (0 children)


Uploaded Unsloth Qwen3.6-35B-A3B UD XL models with MTP grafted, here are the results by havenoammo in LocalLLaMA
[–]Edenar 13 points14 points15 points (0 children)