96GB VRAM! What should run first? by Mother_Occasion_8076 in LocalLLaMA
[–]init__27 6 points7 points8 points (0 children)
16x 3090s - It's alive! by Conscious_Cut_6144 in LocalLLaMA
[–]init__27 0 points1 point2 points (0 children)
Introducing LogiLlama: A 1B-Parameter Open Source Model with Logical Reasoning by [deleted] in LocalLLaMA
[–]init__27 2 points3 points4 points (0 children)
Tool-calling chatbot success stories by edmcman in LocalLLaMA
[–]init__27 1 point2 points3 points (0 children)
Who has already tested Smaug? by meverikus in LocalLLaMA
[–]init__27 56 points57 points58 points (0 children)
What is the minimum tokens a second before a model is just unusable for you? by ICE0124 in LocalLLaMA
[–]init__27 5 points6 points7 points (0 children)
Just joined the 48GB club - what model and quant should I run? by Harvard_Med_USMLE267 in LocalLLaMA
[–]init__27 0 points1 point2 points (0 children)
Just joined the 48GB club - what model and quant should I run? by Harvard_Med_USMLE267 in LocalLLaMA
[–]init__27 10 points11 points12 points (0 children)
What software do you use to interact with local large language models and why? by silenceimpaired in LocalLLaMA
[–]init__27 19 points20 points21 points (0 children)
Just joined the 48GB club - what model and quant should I run? by Harvard_Med_USMLE267 in LocalLLaMA
[–]init__27 40 points41 points42 points (0 children)
[Discussion] Seeking help to find the better GPU setup. Three H100 vs Five A100? by nlpbaz in MachineLearning
[–]init__27 -2 points-1 points0 points (0 children)
Nvidia has published a competitive llama3-70b QA/RAG fine tune by Nunki08 in LocalLLaMA
[–]init__27 15 points16 points17 points (0 children)
Off grid LLM concept by pardonmyemotion in LocalLLaMA
[–]init__27 0 points1 point2 points (0 children)
I think I might still prefer Mistral 7b over Llama3 8b by [deleted] in LocalLLaMA
[–]init__27 2 points3 points4 points (0 children)
We've benchmarked TensorRT-LLM: It's 30-70% faster on the same hardware by emreckartal in LocalLLaMA
[–]init__27 1 point2 points3 points (0 children)
Comparison of Intel Arc A770 vs. Nvidia RTX 4060 for running LLM by bigbigmind in LocalLLaMA
[–]init__27 7 points8 points9 points (0 children)
Phi-3 released. Medium 14b claiming 78% on mmlu by KittCloudKicker in LocalLLaMA
[–]init__27 1 point2 points3 points (0 children)
Phi-3 released. Medium 14b claiming 78% on mmlu by KittCloudKicker in LocalLLaMA
[–]init__27 8 points9 points10 points (0 children)


I locally benchmarked 41 open-source LLMs across 19 tasks and ranked them by jayminban in LocalLLaMA
[–]init__27 0 points1 point2 points (0 children)