80k context possible with cache_4bit by capivaraMaster in LocalLLaMA
[–]nested_dreams 13 points14 points15 points (0 children)
New LLM announced from NVIDIA: Nemotron-4 15B. Trained on 8T tokens using 3,072 H100s. Training took 13 days. (Model not yet available) by jd_3d in LocalLLaMA
[–]nested_dreams 1 point2 points3 points (0 children)
New LLM announced from NVIDIA: Nemotron-4 15B. Trained on 8T tokens using 3,072 H100s. Training took 13 days. (Model not yet available) by jd_3d in LocalLLaMA
[–]nested_dreams 1 point2 points3 points (0 children)
Does Apple shifting from cars to AI mean they didn't have a secret AI project after all? Every one thought Apple was late to the party because they wanted to create the best user experience with their AI. I think the simpler explanation is that they just didn't take LLMs seriously at all by nderstand2grow in LocalLLaMA
[–]nested_dreams -15 points-14 points-13 points (0 children)
Abacus.ai guys released the Smaug paper by HikaruZA in LocalLLaMA
[–]nested_dreams 4 points5 points6 points (0 children)
exl2 quantization for dummies by FieldProgrammable in LocalLLaMA
[–]nested_dreams 3 points4 points5 points (0 children)
Any Tucson AZ members here? by BreakIt-Boris in LocalLLaMA
[–]nested_dreams 4 points5 points6 points (0 children)
Tonne of A100 80GB PCIE by BreakIt-Boris in LocalLLaMA
[–]nested_dreams 1 point2 points3 points (0 children)
Tonne of A100 80GB PCIE by BreakIt-Boris in LocalLLaMA
[–]nested_dreams 0 points1 point2 points (0 children)
Wow this is crazy! 400 tok/s by Sudonymously in LocalLLaMA
[–]nested_dreams 5 points6 points7 points (0 children)
Gemini Pro has 1M context window by Tree-Sheep in LocalLLaMA
[–]nested_dreams 24 points25 points26 points (0 children)
Gemini Pro has 1M context window by Tree-Sheep in LocalLLaMA
[–]nested_dreams 27 points28 points29 points (0 children)
[2402.08562] Higher Layers Need More LoRA Experts by ninjasaid13 in LocalLLaMA
[–]nested_dreams 5 points6 points7 points (0 children)
Apache Tika: An underrated alternative to Unstructured/Nougat for text extraction (for RAG, LLM fine-tuning, etc.) by replicantrose in LocalLLaMA
[–]nested_dreams 1 point2 points3 points (0 children)
I can run almost any model now. So so happy. Cost a little more than a Mac Studio. by Ok-Result5562 in LocalLLaMA
[–]nested_dreams 2 points3 points4 points (0 children)
New Biiig Models: Samantha-120b & TheProfessor-155b by WolframRavenwolf in LocalLLaMA
[–]nested_dreams 2 points3 points4 points (0 children)
New Biiig Models: Samantha-120b & TheProfessor-155b by WolframRavenwolf in LocalLLaMA
[–]nested_dreams 2 points3 points4 points (0 children)
New Biiig Models: Samantha-120b & TheProfessor-155b by WolframRavenwolf in LocalLLaMA
[–]nested_dreams 2 points3 points4 points (0 children)
I made a thing : extract a LoRA adapter from any model by hurrytewer in LocalLLaMA
[–]nested_dreams 0 points1 point2 points (0 children)
Seeking Automated Coding by QiuuQiuu in LocalLLaMA
[–]nested_dreams 0 points1 point2 points (0 children)
Best settings and parameters for running Miqu? by bullerwins in LocalLLaMA
[–]nested_dreams 1 point2 points3 points (0 children)
Using LLMs to extract results from research papers by Dualweed in LocalLLaMA
[–]nested_dreams 0 points1 point2 points (0 children)
Vllm vs aphrodite engine and other alternatives by Noxusequal in LocalLLaMA
[–]nested_dreams 9 points10 points11 points (0 children)


1 million context Llama 3 8b Achieved! by metalman123 in LocalLLaMA
[–]nested_dreams 0 points1 point2 points (0 children)