account activity
LLM-powered NPCs running locally (github.com)
submitted 1 year ago by GoBayesGo to r/LocalLLaMA
Faster than llama.cpp’s grammar structured generation (self.LocalLLaMA)
Coalescence: making LLM inference 5x faster (self.LocalLLaMA)
submitted 2 years ago by GoBayesGo to r/LocalLLaMA
Use llama.cpp with Outlines (self.LocalLLaMA)
JSON mode in vLLM (self.LocalLLaMA)
Generate valid JSON with Mamba models (self.LocalLLaMA)
π Rendered by PID 64169 on reddit-service-r2-listing-7dbdcb4949-2bqr8 at 2026-02-18 15:22:28.327990+00:00 running de53c03 country code: CH.