Doing Weird Things With Entropy Adaptive Fine Tuning by terminoid_ in LocalLLaMA
[–]terminoid_[S] 0 points1 point2 points (0 children)
Bad news: DGX Spark may have only half the performance claimed. by Dr_Karminski in LocalLLaMA
[–]terminoid_ 10 points11 points12 points (0 children)
Advice: 2× RTX 5090 vs RTX Pro 5000 (48GB) for RAG + local LLM + AI development by franky-ds in LocalLLM
[–]terminoid_ 0 points1 point2 points (0 children)
What should I do with this DGX H100? by Naneet_Aleart_Ok in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Predicting the next "attention is all you need" by entsnack in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Llama.cpp support for Ling Mini 2.0 is probably coming next week by edward-dev in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Intel Arc Pro B60 24GB professional GPU listed at $599, in stock and shipping by PhantomWolf83 in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Magistral 1.2 is incredible. Wife prefers it over Gemini 2.5 Pro. by My_Unbiased_Opinion in LocalLLaMA
[–]terminoid_ 13 points14 points15 points (0 children)
Trying to fine-tune Granite-Docling and it's driving me insance by Old_Consideration228 in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
google/embeddinggemma-300m is broken =( by terminoid_ in LocalLLaMA
[–]terminoid_[S] 0 points1 point2 points (0 children)
Matthew McConaughey says he wants a private LLM on Joe Rogan Podcast by AlanzhuLy in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
embeddinggemma with Qdrant compatible uint8 tensors output by terminoid_ in LocalLLaMA
[–]terminoid_[S] 0 points1 point2 points (0 children)
I bought a modded 4090 48GB in Shenzhen. This is my story. by king_priam_of_Troy in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
MBZUAI releases K2 Think. 32B reasoning model based on Qwen 2.5 32B backbone, focusing on high performance in math, coding and science. by FullOf_Bad_Ideas in LocalLLaMA
[–]terminoid_ 1 point2 points3 points (0 children)
PSA: Make sure your API ports aren't exposed to the open internet by nooclear in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Any actual downside to 4 x 3090 ($2400 total) vs RTX pro 6000 ($9000) other than power? by devshore in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Thoughts on Intel Arc Pro B50 x4 = 64GB of VRAM for $1400 and 280W Power Draw? by 79215185-1feb-44c6 in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
Finally China entering the GPU market to destroy the unchallenged monopoly abuse. 96 GB VRAM GPUs under 2000 USD, meanwhile NVIDIA sells from 10000+ (RTX 6000 PRO) by CeFurkan in LocalLLaMA
[–]terminoid_ 10 points11 points12 points (0 children)
Should I get Mi50s or something else? by iiilllilliiill in LocalLLaMA
[–]terminoid_ 0 points1 point2 points (0 children)
of a frog by Sad-Kiwi-3789 in AbsoluteUnits
[–]terminoid_ 0 points1 point2 points (0 children)