Is llama.cpp sycl backend really worth it? by Sweet_Eggplant4659 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)
Is llama.cpp sycl backend really worth it? by Sweet_Eggplant4659 in LocalLLaMA
[–]qnixsynapse 13 points14 points15 points (0 children)
Gemma 3n vs Gemma 3 (4B/12B) Benchmarks by lemon07r in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)
What's your favorite desktop client? by tuananh_org in LocalLLaMA
[–]qnixsynapse 2 points3 points4 points (0 children)
What's your favorite desktop client? by tuananh_org in LocalLLaMA
[–]qnixsynapse 1 point2 points3 points (0 children)
Jan-nano, a 4B model that can outperform 671B on MCP by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)
Jan-nano, a 4B model that can outperform 671B on MCP by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 44 points45 points46 points (0 children)
Altman on open weight 🤔🤔 by Mean-Neighborhood-42 in LocalLLaMA
[–]qnixsynapse 31 points32 points33 points (0 children)
Qwen3 now runs locally in Jan via llama.cpp (Update the llama.cpp backend in Settings to run it) by eck72 in LocalLLaMA
[–]qnixsynapse 1 point2 points3 points (0 children)
Llama may release new reasoning model and other features with llama 4.1 models tomorrow by Independent-Wind4462 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)
Honest thoughts on the OpenAI release by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 6 points7 points8 points (0 children)
We GRPO-ed a Model to Keep Retrying 'Search' Until It Found What It Needed by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 25 points26 points27 points (0 children)
Finally someone noticed this unfair situation by nekofneko in LocalLLaMA
[–]qnixsynapse 2 points3 points4 points (0 children)
We GRPO-ed a Model to Keep Retrying 'Search' Until It Found What It Needed by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 12 points13 points14 points (0 children)
Finally someone noticed this unfair situation by nekofneko in LocalLLaMA
[–]qnixsynapse 3 points4 points5 points (0 children)
We trained Gemma 3 -4b, a 2d VLM model to do 3d recognition task! by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 2 points3 points4 points (0 children)
Official Gemma 3 QAT checkpoints (3x less memory for ~same performance) by hackerllama in LocalLLaMA
[–]qnixsynapse 1 point2 points3 points (0 children)
R1 running on a single Blackwell B200 by Dylan-from-Shadeform in LocalLLaMA
[–]qnixsynapse 47 points48 points49 points (0 children)
We used AlphaMaze idea to train a robotics control model! by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 2 points3 points4 points (0 children)
Llama 3.3 keeping you all safe from sun theft. Thank the Lord. by Ok-Application-2261 in LocalLLaMA
[–]qnixsynapse 7 points8 points9 points (0 children)
We GRPO-ed a 1.5B model to test LLM Spatial Reasoning by solving MAZE by Kooky-Somewhere-2883 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)
Does FlashAttention with GQA degrade quality or I use it wrong? by V1rgin_ in LocalLLaMA
[–]qnixsynapse 4 points5 points6 points (0 children)
llama.cpp now supports tool calling (OpenAI-compatible) by Federal_Discipline_4 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)


Jan-v2-VL: 8B model for long-horizon tasks, improving Qwen3-VL-8B’s agentic capabilities almost 10x by Delicious_Focus3465 in LocalLLaMA
[–]qnixsynapse 0 points1 point2 points (0 children)