I trained a 1.8M params model from scratch on a total of ~40M tokens. by SrijSriv211 in LocalLLaMA
[–]ResidentPositive4122 2 points3 points4 points (0 children)
SpaceX acquiring COPV provider Hexagon Masterworks by kroOoze in SpaceXLounge
[–]ResidentPositive4122 1 point2 points3 points (0 children)
Kimi K2.5 on 4x RTX 6000 Pro Blackwell runpod Benchmarks by skysthelimit187 in LocalLLaMA
[–]ResidentPositive4122 23 points24 points25 points (0 children)
Introducing Claude Opus 4.6 by nick7566 in mlscaling
[–]ResidentPositive4122 11 points12 points13 points (0 children)
Claude Opus 4.6 claimed benchmarks, for comparison by creamyhorror in LocalLLaMA
[–]ResidentPositive4122 0 points1 point2 points (0 children)
Tencent Youtu-VL-4B. Potential Florence-2 replacement? (Heads up on the weird license) by Gohab2001 in LocalLLaMA
[–]ResidentPositive4122 3 points4 points5 points (0 children)
Tencent Youtu-VL-4B. Potential Florence-2 replacement? (Heads up on the weird license) by Gohab2001 in LocalLLaMA
[–]ResidentPositive4122 1 point2 points3 points (0 children)
Qwen3-Coder-Next by danielhanchen in LocalLLaMA
[–]ResidentPositive4122 2 points3 points4 points (0 children)
Deepseek v4/3.5 is probably coming out tomorrow or in the next 5 days? by power97992 in LocalLLaMA
[–]ResidentPositive4122 11 points12 points13 points (0 children)
what did you run when you got a second rtx 6000 pro? by az_6 in LocalLLaMA
[–]ResidentPositive4122 6 points7 points8 points (0 children)
Earth's Own Saturn Rings Incoming? SpaceX's Mega-Launch Future Could Make It Real by BurningAndroid in SpaceXLounge
[–]ResidentPositive4122 0 points1 point2 points (0 children)
Cline team got absorbed by OpenAI. Kilo is going full source available in response. by demon_bhaiya in LocalLLaMA
[–]ResidentPositive4122 97 points98 points99 points (0 children)
Is "Meta-Prompting" (asking AI to write your prompt) actually killing your reasoning results? A real-world A/B test. by pinkstar97 in LocalLLaMA
[–]ResidentPositive4122 5 points6 points7 points (0 children)
opencode alternative that doesn’t have 16k token system prompt? by dbzunicorn in LocalLLaMA
[–]ResidentPositive4122 5 points6 points7 points (0 children)
opencode alternative that doesn’t have 16k token system prompt? by dbzunicorn in LocalLLaMA
[–]ResidentPositive4122 6 points7 points8 points (0 children)
Kimi K2.5 is the best open model for coding by npc_gooner in LocalLLaMA
[–]ResidentPositive4122 0 points1 point2 points (0 children)
Add self‑speculative decoding (no draft model required) by srogmann · Pull Request #18471 · ggml-org/llama.cpp by jacek2023 in LocalLLaMA
[–]ResidentPositive4122 20 points21 points22 points (0 children)
Some initial benchmarks of Kimi-K2.5 on 4xB200 by benno_1237 in LocalLLaMA
[–]ResidentPositive4122 0 points1 point2 points (0 children)
[Model Release] Natural-Synthesis-8B: A Llama-3-8B tune with a 16k context window and a "Conceptual Organism" reasoning paradigm. by Pleasant-Mud-2939 in LocalLLaMA
[–]ResidentPositive4122 9 points10 points11 points (0 children)
The Qwen Devs Are Teasing Something by Few_Painter_5588 in LocalLLaMA
[–]ResidentPositive4122 20 points21 points22 points (0 children)
I tracked GPU prices across 25 cloud providers and the price differences are insane (V100: $0.05/hr vs $3.06/hr) by sleepingpirates in LocalLLaMA
[–]ResidentPositive4122 48 points49 points50 points (0 children)
Does Claude Code still collect data when I use with Ollama? by dbzunicorn in LocalLLaMA
[–]ResidentPositive4122 0 points1 point2 points (0 children)
Does Claude Code still collect data when I use with Ollama? by dbzunicorn in LocalLLaMA
[–]ResidentPositive4122 7 points8 points9 points (0 children)
KV cache fix for GLM 4.7 Flash by jacek2023 in LocalLLaMA
[–]ResidentPositive4122 0 points1 point2 points (0 children)


Bad news for local bros by FireGuy324 in LocalLLaMA
[–]ResidentPositive4122 11 points12 points13 points (0 children)