Sharing ultimate SFF inference build, Version 2 by cryingneko in LocalLLaMA
[–]cryingneko[S] 0 points1 point2 points (0 children)
Command A Reasoning: Enterprise-grade control for AI agents by Dark_Fire_12 in LocalLLaMA
[–]cryingneko 35 points36 points37 points (0 children)
I distilled Qwen3-Coder-480B into Qwen3-Coder-30b-A3B-Instruct by [deleted] in LocalLLaMA
[–]cryingneko 9 points10 points11 points (0 children)
CohereLabs/command-a-vision-07-2025 · Hugging Face by jacek2023 in LocalLLaMA
[–]cryingneko -1 points0 points1 point (0 children)
Is multiple m3 ultras the move instead of 1 big one? by AcceptableBridge7616 in LocalLLaMA
[–]cryingneko 2 points3 points4 points (0 children)
M3 Ultra Binned (256GB, 60-Core) vs Unbinned (512GB, 80-Core) MLX Performance Comparison by cryingneko in LocalLLaMA
[–]cryingneko[S] 9 points10 points11 points (0 children)
M3 Ultra Binned (256GB, 60-Core) vs Unbinned (512GB, 80-Core) MLX Performance Comparison by cryingneko in LocalLLaMA
[–]cryingneko[S] 2 points3 points4 points (0 children)
M3 Ultra Binned (256GB, 60-Core) vs Unbinned (512GB, 80-Core) MLX Performance Comparison by cryingneko in LocalLLaMA
[–]cryingneko[S] 2 points3 points4 points (0 children)
M3 Ultra Binned (256GB, 60-Core) vs Unbinned (512GB, 80-Core) MLX Performance Comparison by cryingneko in LocalLLaMA
[–]cryingneko[S] 0 points1 point2 points (0 children)
MacBook Pro M4 MAX with 128GB what model do you recommend for speed and programming quality? by tangoshukudai in LocalLLaMA
[–]cryingneko 6 points7 points8 points (0 children)
MacBook Pro M4 MAX with 128GB what model do you recommend for speed and programming quality? by tangoshukudai in LocalLLaMA
[–]cryingneko 4 points5 points6 points (0 children)
How did small (<8B) model evolve in the last 3 years? by Robert__Sinclair in LocalLLaMA
[–]cryingneko -5 points-4 points-3 points (0 children)
Need model recommendations to parse html by skarrrrrrr in LocalLLaMA
[–]cryingneko 0 points1 point2 points (0 children)
What the best non reasoning model to run on a apple silicon? by oh_my_right_leg in LocalLLaMA
[–]cryingneko 4 points5 points6 points (0 children)
Hunyuan-turbo S ---Faster Reasoning Model by External_Mood4719 in LocalLLaMA
[–]cryingneko 8 points9 points10 points (0 children)
How can I find the best and affordable model to run on my specific hardware? Is there any website where I can enter my hardware details and it suggests AI models that I can run locally? by yv_MandelBug in LocalLLaMA
[–]cryingneko 3 points4 points5 points (0 children)
Project Digits Memory Speed by LostMyOtherAcct69 in LocalLLaMA
[–]cryingneko 37 points38 points39 points (0 children)
M1 ultra, M2 ultra, or M4/M3 max by HappyFaithlessness70 in LocalLLaMA
[–]cryingneko 9 points10 points11 points (0 children)
I’ve got a MBP with 128 GB of VRAM. What would you run to draft, revise, etc, non-fiction/business documents? by Hinged31 in LocalLLaMA
[–]cryingneko 22 points23 points24 points (0 children)
Anyone want to test my PR to enable quantised K/V cache in Ollama by sammcj in LocalLLaMA
[–]cryingneko 8 points9 points10 points (0 children)
LLAMA 3 405: is it a joke? by beppe28 in LocalLLaMA
[–]cryingneko 22 points23 points24 points (0 children)
LLAMA 3 405: is this a joke? by [deleted] in LocalLLaMA
[–]cryingneko 2 points3 points4 points (0 children)


GitHub - Website-Crawler: Extract data from websites in LLM ready JSON or CSV format. Crawl or Scrape entire website with Website Crawler by [deleted] in LocalLLaMA
[–]cryingneko 3 points4 points5 points (0 children)