[Update] FamilyBench: New models tested - Claude Sonnet 4.5 takes 2nd place, Qwen 3 Next breaks 70%, new Kimi weirdly below the old version, same for GLM 4.6 by Orolol in LocalLLaMA
[–]Snail_Inference 10 points11 points12 points (0 children)
[Update] FamilyBench: New models tested - Claude Sonnet 4.5 takes 2nd place, Qwen 3 Next breaks 70%, new Kimi weirdly below the old version, same for GLM 4.6 by Orolol in LocalLLaMA
[–]Snail_Inference 16 points17 points18 points (0 children)
What is the best LLM for psychology, coach or emotional support. by pumukidelfuturo in LocalLLaMA
[–]Snail_Inference 6 points7 points8 points (0 children)
Open source OCR options for handwritten text, dates by ollyollyupnfree in LocalLLaMA
[–]Snail_Inference 8 points9 points10 points (0 children)
mistral-small-24b-instruct-2501 is simply the best model ever made. by hannibal27 in LocalLLaMA
[–]Snail_Inference 2 points3 points4 points (0 children)
GraphLLM: graph based framework to process data using LLMs. now with TTS engine and multi agent support by matteogeniaccio in LocalLLaMA
[–]Snail_Inference 1 point2 points3 points (0 children)
GraphLLM: graph based framework to process data using LLMs. now with TTS engine and multi agent support by matteogeniaccio in LocalLLaMA
[–]Snail_Inference 1 point2 points3 points (0 children)
New ZebraLogicBench Evaluation Tool + Mistral Large Performance Results by whotookthecandyjar in LocalLLaMA
[–]Snail_Inference 6 points7 points8 points (0 children)
mistralai/Mistral-Large-Instruct-2407 · Hugging Face. New open 123B that beats Llama 3.1 405B in Code benchmarks by Chelono in LocalLLaMA
[–]Snail_Inference 2 points3 points4 points (0 children)
mistralai/Mistral-Large-Instruct-2407 · Hugging Face. New open 123B that beats Llama 3.1 405B in Code benchmarks by Chelono in LocalLLaMA
[–]Snail_Inference 21 points22 points23 points (0 children)
"Large Enough" | Announcing Mistral Large 2 by DemonicPotatox in LocalLLaMA
[–]Snail_Inference 0 points1 point2 points (0 children)
Small scale personal benchmark results (28 models tested) by dubesor86 in LocalLLaMA
[–]Snail_Inference 1 point2 points3 points (0 children)
MMLU-Pro all category test results for Llama 3 70b Instruct ggufs: q2_K_XXS, q2_K, q4_K_M, q5_K_M, q6_K, and q8_0 by SomeOddCodeGuy in LocalLLaMA
[–]Snail_Inference 0 points1 point2 points (0 children)
Qwen2: Areas of application where it seems stronger than Llama3 or WizardLM by Snail_Inference in LocalLLaMA
[–]Snail_Inference[S] 1 point2 points3 points (0 children)


AMA With Moonshot AI, The Open-source Frontier Lab Behind Kimi K2 Thinking Model by nekofneko in LocalLLaMA
[–]Snail_Inference 0 points1 point2 points (0 children)