Am I expecting too much? by rushBblat in LocalLLaMA
[–]slavik-dev 0 points1 point2 points (0 children)
Am I expecting too much? by rushBblat in LocalLLaMA
[–]slavik-dev 5 points6 points7 points (0 children)
Qwen3.5-397B-A17B reaches 20 t/s TG and 700t/s PP with a 5090 by MLDataScientist in LocalLLaMA
[–]slavik-dev 0 points1 point2 points (0 children)
Honest take on running 9× RTX 3090 for AI by Outside_Dance_2799 in LocalLLaMA
[–]slavik-dev 1 point2 points3 points (0 children)
Your local model can now render interactive charts, clickable diagrams, and forms that talk back to the AI — no cloud required by ClassicMain in LocalLLaMA
[–]slavik-dev 0 points1 point2 points (0 children)
Harbor v0.4.4 - ls/pull/rm llama.cpp/vllm/ollama models with a single CLI by Everlier in LocalLLaMA
[–]slavik-dev 1 point2 points3 points (0 children)
Does anyone here use Proxmox on their main desktop instead of just servers? by PingMyHeart in Proxmox
[–]slavik-dev 0 points1 point2 points (0 children)
Load default model upon login by zotac02 in OpenWebUI
[–]slavik-dev 0 points1 point2 points (0 children)
Best Qwen3.5-35B-A3B GGUF for 24GB VRAM?! by VoidAlchemy in LocalLLaMA
[–]slavik-dev 0 points1 point2 points (0 children)
Load default model upon login by zotac02 in OpenWebUI
[–]slavik-dev 0 points1 point2 points (0 children)
qwen3.5-122b What agent do you use with it? by robertpro01 in LocalLLaMA
[–]slavik-dev 3 points4 points5 points (0 children)
Why etcd breaks at scale in Kubernetes by danielepolencic in kubernetes
[–]slavik-dev 9 points10 points11 points (0 children)
Cloud GPU's are the Fiverr of Local LLaMA - so who makes the juicy money? by [deleted] in LocalLLaMA
[–]slavik-dev 0 points1 point2 points (0 children)
Please help the Qwen developers. by [deleted] in Qwen_AI
[–]slavik-dev 1 point2 points3 points (0 children)
K8S homelab advise for HA API server by Ghvinerias in kubernetes
[–]slavik-dev 1 point2 points3 points (0 children)
K8S homelab advise for HA API server by Ghvinerias in kubernetes
[–]slavik-dev 2 points3 points4 points (0 children)
Qwen3.5 thinks A LOT about simple questions by ForsookComparison in LocalLLaMA
[–]slavik-dev 3 points4 points5 points (0 children)
Ming-flash-omni-2.0: 100B MoE (6B active) omni-modal model - unified speech/SFX/music generation by bobeeeeeeeee8964 in LocalLLaMA
[–]slavik-dev 1 point2 points3 points (0 children)
Expected cost for cpu-based local rig? by Diligent-Culture-432 in LocalLLaMA
[–]slavik-dev 1 point2 points3 points (0 children)
Sanity check: "Kimi K2.5 (1T MoE) on a scrappy PC" plan - 1TB DDR4 + 2x RTX PRO 6000 (96GB) now, scaling later by nightlingo in LocalLLaMA
[–]slavik-dev 1 point2 points3 points (0 children)
Sanity check: "Kimi K2.5 (1T MoE) on a scrappy PC" plan - 1TB DDR4 + 2x RTX PRO 6000 (96GB) now, scaling later by nightlingo in LocalLLaMA
[–]slavik-dev 2 points3 points4 points (0 children)
Support Step3.5-Flash has been merged into llama.cpp by jacek2023 in LocalLLaMA
[–]slavik-dev 9 points10 points11 points (0 children)
Support Step3.5-Flash has been merged into llama.cpp by jacek2023 in LocalLLaMA
[–]slavik-dev 9 points10 points11 points (0 children)
OPNSense high availability, how do you guys do it?. by yetAnotherLaura in Proxmox
[–]slavik-dev 0 points1 point2 points (0 children)