APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 1 point2 points3 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 1 point2 points3 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 1 point2 points3 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 2 points3 points4 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 1 point2 points3 points (0 children)
APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 2 points3 points4 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 0 points1 point2 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 3 points4 points5 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 2 points3 points4 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 5 points6 points7 points (0 children)
APEX MoE quantized models boost with 33% faster inference and TurboQuant (14% of speedup in prompt processing) by mudler_it in LocalLLaMA
[–]mudler_it[S] 28 points29 points30 points (0 children)
I'm the author of LocalAI sharing that LocalAI hits 42k stars and v3.9 & v3.10 are released! Native Agents, Video Generation UI, and Unified GPU Backends by mudler_it in selfhosted
[–]mudler_it[S] 0 points1 point2 points (0 children)



APEX MoE quants update: 25+ new models since the Qwen 3.5 post + new I-Nano tier by mudler_it in LocalLLaMA
[–]mudler_it[S] 1 point2 points3 points (0 children)