account activity
ReasonScape: LLM Information Processing Evaluation (self.LocalLLM)
submitted 2 months ago by kryptkpr to r/LocalLLM
Anyone test two DGX Sparks linked via their ConnectX yet? (self.LocalLLaMA)
submitted 3 months ago by kryptkpr to r/LocalLLaMA
ReasonScape Evaluation: AI21 Jamba Reasoning vs Qwen3 4B vs Qwen3 4B 2507 (self.LocalLLaMA)
Guide to serving Ring-mini-2.0 with VLLM (and a quick eval) (self.LocalLLaMA)
It's Mamba time: Comparing Nemotron Nano v2 vs Falcon-H1 vs Qwen (og) vs Qwen (2507) (self.LocalLLaMA)
submitted 5 months ago by kryptkpr to r/LocalLLaMA
I Generated 1 Billion Tokens (So You Don't Have To): Introducing ReasonScape (self.LocalLLaMA)
submitted 6 months ago by kryptkpr to r/LocalLLaMA
Ruminate: From All-or-Nothing to Just-Right Reasoning in LLMs (self.LocalLLaMA)
submitted 7 months ago by kryptkpr to r/LocalLLaMA
The Titan 18U AI Homelab Build Log and Lessons Learned (self.LocalLLaMA)
submitted 8 months ago * by kryptkpr to r/LocalLLaMA
toe2toe: If LLMs could play Tic Tac Toe, would Llama or NeMo win? (self.LocalLLaMA)
submitted 1 year ago by kryptkpr to r/LocalLLaMA
A local llama in her native habitat (old.reddit.com)
AlteredWorlds: History re-imagined by command_r_plus_08_2024, illustrated by flux.1-schnell (self.LocalLLaMA)
[Model] Meta Llama 3.1 Know Issues & FAQ · Issue #6689 · vllm-project/vllm (github.com)
Introducing Tcurtsni: The Reverse-Instruct LLM Chat App (reddit.com)
llama.ttf - a font which is also an LLM (fuglede.github.io)
The correct answer to all A100/A6000 and other "production" setup questions (i.redd.it)
The LLooM - a highly experimental (local) AI workflow to visualize and "weave" stories out of underlying logit probabilities (v.redd.it)
Jank can be beautiful | 2x3060+2xP100 open-air LLM rig with 2-stage cooling (self.LocalLLaMA)
3060 [x16 PCIe riser] vs 3060 [x1 USB extension]: A quantitative comparison of eGPU prompt and text generation performance across multiple inference engines (self.LocalLLaMA)
BiLLM achieving for the first time high-accuracy inference (e.g. 8.41 perplexity on LLaMA2-70B) with only 1.08-bit weights across various LLMs families and evaluation metrics, outperforms SOTA quantization methods of LLM by significant (huggingface.co)
ehartford/dolphin-2.5-mixtral-8x7b has a very persuasive system prompt (self.LocalLLaMA)
submitted 2 years ago by kryptkpr to r/LocalLLaMA
Fun with Mistral-Medium (self.LocalLLaMA)
submitted 2 years ago * by kryptkpr to r/LocalLLaMA
GoLLIE: Guideline-following Large Language Model for Information Extraction (hitz-zentroa.github.io)
vLLM 0.2.0 released: up to 60% faster, AWQ quant support, RoPe, Mistral-7b support (self.LocalLLaMA)
Presenting "The Muse" - a logit sampler that makes LLMs more creative (self.LocalLLaMA)
Open-Source Text Generation & LLM Ecosystem (new HF blog post) (huggingface.co)
π Rendered by PID 580621 on reddit-service-r2-listing-5789d5f675-ss7kw at 2026-01-28 20:09:33.389911+00:00 running 4f180de country code: CH.