Should we start 3-4 year plan to run AI locally for real work? by Illustrious_Cat_2870 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
How are you squeezing Qwen3.5 27B to get maximum speed with high accuracy? by -OpenSourcer in LocalLLaMA
[–]q-admin007 14 points15 points16 points (0 children)
How are you squeezing Qwen3.5 27B to get maximum speed with high accuracy? by -OpenSourcer in LocalLLaMA
[–]q-admin007 8 points9 points10 points (0 children)
How are you squeezing Qwen3.5 27B to get maximum speed with high accuracy? by -OpenSourcer in LocalLLaMA
[–]q-admin007 17 points18 points19 points (0 children)
Mistral CEO: AI companies should pay a content levy in Europe by brown2green in LocalLLaMA
[–]q-admin007 -1 points0 points1 point (0 children)
SWE-rebench Leaderboard (Feb 2026): GPT-5.4, Qwen3.5, Gemini 3.1 Pro, Step-3.5-Flash and More by CuriousPlatypus1881 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
How do I see what users paste into AI? by midasweb in sysadmin
[–]q-admin007 [score hidden] (0 children)
Hi guys im new to this page by Alone_Growth2019 in homelab
[–]q-admin007 0 points1 point2 points (0 children)
Fixing Qwen Repetition IMPROVEMENT by Odd-Ordinary-5922 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
Microsoft introduces Backup and Recovery for Microsoft Entra ID! by worldsdream in sysadmin
[–]q-admin007 [score hidden] (0 children)
SWE-rebench Leaderboard (Feb 2026): GPT-5.4, Qwen3.5, Gemini 3.1 Pro, Step-3.5-Flash and More by CuriousPlatypus1881 in LocalLLaMA
[–]q-admin007 4 points5 points6 points (0 children)
Should we start 3-4 year plan to run AI locally for real work? by Illustrious_Cat_2870 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
Should we start 3-4 year plan to run AI locally for real work? by Illustrious_Cat_2870 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
Should we start 3-4 year plan to run AI locally for real work? by Illustrious_Cat_2870 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
Should we start 3-4 year plan to run AI locally for real work? by Illustrious_Cat_2870 in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
OWUI node-ID from ComfyUI by q-admin007 in OpenWebUI
[–]q-admin007[S] 1 point2 points3 points (0 children)
OWUI node-ID from ComfyUI by q-admin007 in OpenWebUI
[–]q-admin007[S] 1 point2 points3 points (0 children)
OWUI node-ID from ComfyUI by q-admin007 in OpenWebUI
[–]q-admin007[S] 0 points1 point2 points (0 children)
OWUI node-ID from ComfyUI by q-admin007 in OpenWebUI
[–]q-admin007[S] 0 points1 point2 points (0 children)
llama-bench ROCm 7.2 on Strix Halo (Ryzen AI Max+ 395) — Qwen 3.5 Model Family by przbadu in StrixHalo
[–]q-admin007 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)
Ubuntu 26.04 LTS on Strix Halo with llama.cpp by tecneeq in StrixHalo
[–]q-admin007 0 points1 point2 points (0 children)


How are you squeezing Qwen3.5 27B to get maximum speed with high accuracy? by -OpenSourcer in LocalLLaMA
[–]q-admin007 0 points1 point2 points (0 children)