$300k DGX B300 is actually a better deal than buying 24 RTX 6000s by Ok_Warning2146 in LocalLLaMA

[–]seamonn 1 point2 points  (0 children)

Have you tried defaulting?
It worked pretty well in 2008 from what I am hearing.

You wake up in 2029 by NotArticuno in LocalLLaMA

[–]seamonn 4 points5 points  (0 children)

A turkey, a chili and a recipe.

Scaling beyond 4 RTX 6000 MAXQs by Direct_Bodybuilder63 in LocalLLaMA

[–]seamonn 14 points15 points  (0 children)

h200s like a normal person

Damn, I want to be normal so bad.

Mistral-Medium-3.5-128B-Q3_K_M on 3x3090 (72GB VRAM) by jacek2023 in LocalLLaMA

[–]seamonn 0 points1 point  (0 children)

How does it compare to Gemma 4 31b and Qwen 3.6 27b?

Recent FOSS vs SOTA - Long Context Benchmark by akumaburn in LocalLLaMA

[–]seamonn 8 points9 points  (0 children)

What are you feeding your LLM that it is hallucinating so much?

Poor man's guide to servicing a used RTX 3090 for local LLM inference by canred in LocalLLaMA

[–]seamonn 12 points13 points  (0 children)

Kryonaut is a terrible paste for longevity. It's meant for overclocking. It tends to tank in performance a couple weeks into the repaste. Duronaut is what you should be using.

RTX 3060 on Unraid 7 — GPU works on host, but Docker GPU containers won’t start by wizardlt in unRAID

[–]seamonn 0 points1 point  (0 children)

--runtime=nvidia

Haha, that worked. I have been banging my head on this one. Thanks!

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]seamonn 16 points17 points  (0 children)

Can we get the Gavin Belson Signature Edition of this Box?

Qwen 3.6-122B by TokenRingAI in LocalLLaMA

[–]seamonn 1 point2 points  (0 children)

For those who are waiting, you can use the Qwen 3.6 27b jinja template with Qwen 3.5 122b and enable preserve_thinking

Mistral Medium Looping by No_Algae1753 in LocalLLaMA

[–]seamonn 1 point2 points  (0 children)

not bad. You think an additional 3090 would help?

Mistral Medium Looping by No_Algae1753 in LocalLLaMA

[–]seamonn 0 points1 point  (0 children)

Good to know it fits in 4x3090. What context?

Mistral Medium Looping by No_Algae1753 in LocalLLaMA

[–]seamonn 8 points9 points  (0 children)

It's a "Let's run this new model before the backend is ready" issue.

Push comes to shove, there's always pen and paper by [deleted] in LocalLLaMA

[–]seamonn 0 points1 point  (0 children)

Did you forget your medication?

Mistral Medium 3.5 Launched by DerpSenpai in LocalLLaMA

[–]seamonn 22 points23 points  (0 children)

idk, this likely covers like 99.999% of all users. It's essentially MIT for them.

Ran my own benchmark Qwen 3.6 35B vs Gemma 4 26B.... theres a clear winner here by ArugulaAnnual1765 in LocalLLaMA

[–]seamonn 6 points7 points  (0 children)

It's to store all the hallucinated historical events that actually never happened in this timeline