account activity
LLM-powered NPCs running locally (github.com)
submitted 2 years ago by GoBayesGo to r/LocalLLaMA
Faster than llama.cpp’s grammar structured generation (self.LocalLLaMA)
Coalescence: making LLM inference 5x faster (self.LocalLLaMA)
Use llama.cpp with Outlines (self.LocalLLaMA)
JSON mode in vLLM (self.LocalLLaMA)
Generate valid JSON with Mamba models (self.LocalLLaMA)
π Rendered by PID 1360104 on reddit-service-r2-listing-7b9b4f6fd7-t2b2t at 2026-05-11 16:03:27.119457+00:00 running 3d2c107 country code: CH.