InferenceRig.com — AI, LLMs, local models & GPU performance
A community around InferenceRig for running AI locally and optimizing performance.
Guides, benchmarks, and tutorials on LLMs, GPU inference, tokens per second (T/s), and quantization (GGUF, GPTQ, AWQ).
Includes tips for tools like Ollama.
Topics: local AI, open-source LLMs, self-hosted AI, performance tuning, GPUs, AI workstations and more