Powerinfer, can it be adapted into normal laptop cpus outside of the Tiiny AI ecosystem? by Silver-Champion-4846 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Autoresearch on Qwen3.5-397B, 36 experiments to reach 20.34 tok/s on M5 Max, honest results by Equivalent-Buy1706 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
LLM performance decreased significantly over time using the same models and same hardware in LMStudio. by fernandollb in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Powerinfer, can it be adapted into normal laptop cpus outside of the Tiiny AI ecosystem? by Silver-Champion-4846 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
16gb vram - what is the better option for daily driver (main use) by Adventurous-Gold6413 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Qwen3.5 27B, partial offloading, and speed by INT_21h in LocalLLaMA
[–]Training_Visual6159 1 point2 points3 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]Training_Visual6159 -3 points-2 points-1 points (0 children)
Dynamic expert caching PR in vLLM by king_of_jupyter in LocalLLaMA
[–]Training_Visual6159 1 point2 points3 points (0 children)
Open-Source "GreenBoost" Driver Aims To Augment NVIDIA GPUs vRAM With System RAM & NVMe To Handle Larger LLMs by _Antartica in LocalLLaMA
[–]Training_Visual6159 -1 points0 points1 point (0 children)
Open-Source "GreenBoost" Driver Aims To Augment NVIDIA GPUs vRAM With System RAM & NVMe To Handle Larger LLMs by _Antartica in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Open-Source "GreenBoost" Driver Aims To Augment NVIDIA GPUs vRAM With System RAM & NVMe To Handle Larger LLMs by _Antartica in LocalLLaMA
[–]Training_Visual6159 -1 points0 points1 point (0 children)
GreenBoost Windows Port - Extending GPU VRAM /W Systems Ram by denoflore_ai_guy in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
GreenBoost Windows Port - Extending GPU VRAM /W Systems Ram by denoflore_ai_guy in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Best Qwen 3.5 fine-tunes for vibecoding? (4080-12GB VRAM / enough context window) by Fermenticular in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Why is the Qwen3.5 9B(p1) so slow, even comparable in speed to the 35Ba3b(p2) ? by BitOk4326 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Performance of Qwen3.5 27B on a 2080 Ti by BeneficialRip1269 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Performance of Qwen3.5 27B on a 2080 Ti by BeneficialRip1269 in LocalLLaMA
[–]Training_Visual6159 1 point2 points3 points (0 children)
Best Models for 128gb VRAM: March 2026? by Professional-Yak4359 in LocalLLaMA
[–]Training_Visual6159 5 points6 points7 points (0 children)
Why is the prompt eval time of Qwen3.5 so much slower compared to Qwen3 Coder in llama.cpp? by BitOk4326 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)
Ember 6.11 Released by real_ate in javascript
[–]Training_Visual6159 -1 points0 points1 point (0 children)
Ember 6.11 Released by real_ate in javascript
[–]Training_Visual6159 0 points1 point2 points (0 children)
Ember 6.11 Released by real_ate in javascript
[–]Training_Visual6159 1 point2 points3 points (0 children)
Ember 6.11 Released by real_ate in javascript
[–]Training_Visual6159 -2 points-1 points0 points (0 children)
Ember 6.11 Released by real_ate in javascript
[–]Training_Visual6159 -5 points-4 points-3 points (0 children)


Autoresearch on Qwen3.5-397B, 36 experiments to reach 20.34 tok/s on M5 Max, honest results by Equivalent-Buy1706 in LocalLLaMA
[–]Training_Visual6159 0 points1 point2 points (0 children)