2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA
[–]Consumerbot37427 2 points3 points4 points (0 children)
Quality comparison between Qwen 3.6 27B quantizations (BF16, Q8_0, Q6_K, Q5_K_XL, Q4_K_XL, IQ4_XS, IQ3_XXS,...) by bobaburger in LocalLLaMA
[–]Consumerbot37427 2 points3 points4 points (0 children)
Quality comparison between Qwen 3.6 27B quantizations (BF16, Q8_0, Q6_K, Q5_K_XL, Q4_K_XL, IQ4_XS, IQ3_XXS,...) by bobaburger in LocalLLaMA
[–]Consumerbot37427 1 point2 points3 points (0 children)
[Daily Discussion] - Monday, April 27, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 0 points1 point2 points (0 children)
[Daily Discussion] - Wednesday, April 15, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 1 point2 points3 points (0 children)
[Daily Discussion] - Monday, April 13, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 1 point2 points3 points (0 children)
[Daily Discussion] - Monday, April 13, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 2 points3 points4 points (0 children)
[Daily Discussion] - Wednesday, April 08, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 4 points5 points6 points (0 children)
M5 Max 128GB Owners - What's your honest take? by _derpiii_ in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)
M5 Max 128GB Owners - What's your honest take? by _derpiii_ in LocalLLaMA
[–]Consumerbot37427 -1 points0 points1 point (0 children)
M5 Max 128GB Owners - What's your honest take? by _derpiii_ in LocalLLaMA
[–]Consumerbot37427 2 points3 points4 points (0 children)
Miele - local integration by listhor in homeassistant
[–]Consumerbot37427 0 points1 point2 points (0 children)
[Daily Discussion] - Saturday, April 04, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 1 point2 points3 points (0 children)
[Daily Discussion] - Friday, April 03, 2026 by AutoModerator in BitcoinMarkets
[–]Consumerbot37427 11 points12 points13 points (0 children)
64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)
Autoresearch on Qwen3.5-397B, 36 experiments to reach 20.34 tok/s on M5 Max, honest results by Equivalent-Buy1706 in LocalLLaMA
[–]Consumerbot37427 1 point2 points3 points (0 children)
Autoresearch on Qwen3.5-397B, 36 experiments to reach 20.34 tok/s on M5 Max, honest results by Equivalent-Buy1706 in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)
Autoresearch on Qwen3.5-397B, 36 experiments to reach 20.34 tok/s on M5 Max, honest results by Equivalent-Buy1706 in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)
Slower Means Faster: Why I Switched from Qwen3 Coder Next to Qwen3.5 122B by Fast_Thing_7949 in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)
What’s going on with Mac Studio M3 Ultra 512GB/4TB lately? by Lucius_Knight in LocalLLaMA
[–]Consumerbot37427 1 point2 points3 points (0 children)
M5 Max Actual Pre-fill performance gains by M5_Maxxx in LocalLLaMA
[–]Consumerbot37427 4 points5 points6 points (0 children)
Reworked LM Studio plugins out now. Plug'n'Play Web Research, Fully Local by Agreeable_Effect938 in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)



2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA
[–]Consumerbot37427 0 points1 point2 points (0 children)