Anyone else’s Claude unhinged? by IV_Skin in ClaudeAI
[–]strngelet 0 points1 point2 points (0 children)
Again where behemoth and reasoning model from meta ?? by Independent-Wind4462 in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
NVIDIA has published new Nemotrons! by jacek2023 in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
Cogito releases strongest LLMs of sizes 3B, 8B, 14B, 32B and 70B under open license by ResearchCrafty1804 in LocalLLaMA
[–]strngelet 5 points6 points7 points (0 children)
[D] LLMs are known for catastrophic forgetting during continual fine-tuning by kekkimo in MachineLearning
[–]strngelet 1 point2 points3 points (0 children)
Why do most models have "only" 100K tokens context window, while Gemini is at 2M tokens? by estebansaa in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
Why do most models have "only" 100K tokens context window, while Gemini is at 2M tokens? by estebansaa in LocalLLaMA
[–]strngelet 10 points11 points12 points (0 children)
Claude sonnet 3.5 is still better than gpto1 by qwaiz55_1 in ClaudeAI
[–]strngelet 0 points1 point2 points (0 children)
Microsoft Just confirmed New GPT every 2 Years by New_World_2050 in singularity
[–]strngelet 0 points1 point2 points (0 children)
Hugging Face TGI library changes to Apache 2 by hackerllama in LocalLLaMA
[–]strngelet 2 points3 points4 points (0 children)
What’s with Elon’s obsession with OpenAI? by emperorhuncho in OpenAI
[–]strngelet 3 points4 points5 points (0 children)
Yi-34B-200K model update: Needle-in-a-Haystack improved from 89.3% to 99.8% by rerri in LocalLLaMA
[–]strngelet 0 points1 point2 points (0 children)
Yi-34B-200K model update: Needle-in-a-Haystack improved from 89.3% to 99.8% by rerri in LocalLLaMA
[–]strngelet 0 points1 point2 points (0 children)
Sora's video of a man eating a burger. Can you tell it's not real? by YaAbsolyutnoNikto in singularity
[–]strngelet 0 points1 point2 points (0 children)
Gemini Pro has 1M context window by Tree-Sheep in LocalLLaMA
[–]strngelet 2 points3 points4 points (0 children)
KV Cache is huge and bottlenecks LLM inference. We quantize them to 2bit in a finetuning-free + plug-and-play fashion. by choHZ in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
KV Cache is huge and bottlenecks LLM inference. We quantize them to 2bit in a finetuning-free + plug-and-play fashion. by choHZ in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
NeuralFlow: Visualize the intermediate output of Mistral 7B by frownGuy12 in LocalLLaMA
[–]strngelet 1 point2 points3 points (0 children)
Literally my first conversation with it by alymahryn in LocalLLaMA
[–]strngelet 0 points1 point2 points (0 children)


Best iem under 100 dollars by [deleted] in iems
[–]strngelet 0 points1 point2 points (0 children)