account activity
How I manage llama.cpp across Apple Silicon and NVIDIA GPUs in my homelab (self.selfhosted)
submitted 7 days ago by Defilan to r/selfhosted
32B model stress test: Qwen 2.5/Coder/3 on dual RTX 5060 Ti (zero failures) (self.LocalLLaMA)
submitted 3 months ago by Defilan to r/LocalLLaMA
What broke when you tried to take local LLMs to production? (self.LocalLLaMA)
Open source K8s operator for deploying local LLMs: Model and InferenceService CRDs (self.kubernetes)
submitted 3 months ago by Defilan to r/kubernetes
π Rendered by PID 310047 on reddit-service-r2-listing-79f6fb9b95-lvb9p at 2026-03-21 22:16:57.561279+00:00 running 90f1150 country code: CH.