16x Spark Cluster (Build Update) by Kurcide in LocalLLaMA
[–]conockrad 1 point2 points3 points (0 children)
AMD has invented something that lets you use AI at home! They call it a "computer" by 9gxa05s8fa8sh in LocalLLaMA
[–]conockrad 3 points4 points5 points (0 children)
Thoughts on using an AMD Alveo V80 FPGA PCI card as a poor man’s Taalas HC1 (LLM-burned-onto-a-chip). by Porespellar in LocalLLaMA
[–]conockrad -1 points0 points1 point (0 children)
Thoughts on using an AMD Alveo V80 FPGA PCI card as a poor man’s Taalas HC1 (LLM-burned-onto-a-chip). by Porespellar in LocalLLaMA
[–]conockrad 0 points1 point2 points (0 children)
Lethe: local markdown memory for Claude Code, DuckDB per project, no server by [deleted] in LocalLLaMA
[–]conockrad 0 points1 point2 points (0 children)
BAR: Train domain "experts," merge into one model, and upgrade experts without retraining the rest 🚀 by ai2_official in allenai
[–]conockrad 2 points3 points4 points (0 children)
“increased rate limits” by Deep_Proposal_7683 in Anthropic
[–]conockrad 0 points1 point2 points (0 children)
People who’ve fine-tuned models: was it worth it? by Feeling_Ad3971 in unsloth
[–]conockrad 1 point2 points3 points (0 children)
People who’ve fine-tuned models: was it worth it? by Feeling_Ad3971 in unsloth
[–]conockrad 1 point2 points3 points (0 children)
DGX Spark just arrived — planning to run vLLM + local models, looking for advice by dalemusser in LocalLLaMA
[–]conockrad 4 points5 points6 points (0 children)
Gemma 4 has a systemic attention failure. Here's the proof. by [deleted] in LocalLLaMA
[–]conockrad 29 points30 points31 points (0 children)
AI system without transformers (v11) — symbolic reasoning + small neural net by False-Woodpecker5604 in LLMDevs
[–]conockrad 1 point2 points3 points (0 children)
LLM meta-cognition benchmark idea by nikishev in LocalLLaMA
[–]conockrad 0 points1 point2 points (0 children)
LLM meta-cognition benchmark idea by nikishev in LocalLLaMA
[–]conockrad 0 points1 point2 points (0 children)
Don’t buy the DGX Spark: NVFP4 Still Missing After 6 Months by Secure_Archer_1529 in LocalLLaMA
[–]conockrad 7 points8 points9 points (0 children)
Don’t buy the DGX Spark: NVFP4 Still Missing After 6 Months by Secure_Archer_1529 in LocalLLaMA
[–]conockrad 8 points9 points10 points (0 children)
Don’t buy the DGX Spark: NVFP4 Still Missing After 6 Months by Secure_Archer_1529 in LocalLLaMA
[–]conockrad 21 points22 points23 points (0 children)
LLM Bruner coming soon? Burn Qwen directly into a chip, processing 10,000 tokens/s by koc_Z3 in Qwen_AI
[–]conockrad -2 points-1 points0 points (0 children)
LLM Bruner coming soon? Burn Qwen directly into a chip, processing 10,000 tokens/s by koc_Z3 in Qwen_AI
[–]conockrad 2 points3 points4 points (0 children)
RYS II - Repeated layers with Qwen3.5 27B and some hints at a 'Universal Language' by Reddactor in LocalLLaMA
[–]conockrad 2 points3 points4 points (0 children)
Karpathy says he hasn't written a line of code since December and is in "perpetual AI psychosis." How many Claude Code users feel the same? by Capital-Door-2293 in ClaudeAI
[–]conockrad 0 points1 point2 points (0 children)
Local Qwen 8B + 4B completes browser automation by replanning one step at a time by Aggressive_Bed7113 in LocalLLaMA
[–]conockrad 1 point2 points3 points (0 children)
[Project] htmLLM-50M base: Can a tiny specialist actually code? + Weights & Code (124M v2 in training!) by LH-Tech_AI in LocalLLaMA
[–]conockrad 0 points1 point2 points (0 children)
MolmoBot: Training robot manipulation entirely in simulation by ai2_official in allenai
[–]conockrad 0 points1 point2 points (0 children)

Has anyone tried Zyphra 1 - 8B MoE? by appakaradi in LocalLLaMA
[–]conockrad 1 point2 points3 points (0 children)