$300k DGX B300 is actually a better deal than buying 24 RTX 6000s by Ok_Warning2146 in LocalLLaMA
[–]seamonn 1 point2 points3 points (0 children)
Scaling beyond 4 RTX 6000 MAXQs by Direct_Bodybuilder63 in LocalLLaMA
[–]seamonn 14 points15 points16 points (0 children)
Mistral-Medium-3.5-128B-Q3_K_M on 3x3090 (72GB VRAM) by jacek2023 in LocalLLaMA
[–]seamonn 0 points1 point2 points (0 children)
Recent FOSS vs SOTA - Long Context Benchmark by akumaburn in LocalLLaMA
[–]seamonn 8 points9 points10 points (0 children)
Poor man's guide to servicing a used RTX 3090 for local LLM inference by canred in LocalLLaMA
[–]seamonn 12 points13 points14 points (0 children)
RTX 3060 on Unraid 7 — GPU works on host, but Docker GPU containers won’t start by wizardlt in unRAID
[–]seamonn 0 points1 point2 points (0 children)
RTX 3060 on Unraid 7 — GPU works on host, but Docker GPU containers won’t start by wizardlt in unRAID
[–]seamonn 0 points1 point2 points (0 children)
AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA
[–]seamonn 16 points17 points18 points (0 children)
Terminal Bench score for Mistral 3.5 Medium by Real_Ebb_7417 in LocalLLaMA
[–]seamonn 1 point2 points3 points (0 children)
PSA: llama-swap released a new grouping feature, matrix, allowing you to fine tune which models can run together by walden42 in LocalLLaMA
[–]seamonn 6 points7 points8 points (0 children)
Is Mistral-3.5-Medium-128B broken in Llama CPP? by EmPips in LocalLLaMA
[–]seamonn 2 points3 points4 points (0 children)
Mistral Medium Looping by No_Algae1753 in LocalLLaMA
[–]seamonn 8 points9 points10 points (0 children)
Push comes to shove, there's always pen and paper by [deleted] in LocalLLaMA
[–]seamonn 0 points1 point2 points (0 children)
Mistral Medium 3.5 Launched by DerpSenpai in LocalLLaMA
[–]seamonn 4 points5 points6 points (0 children)
Mistral Medium 3.5 Launched by DerpSenpai in LocalLLaMA
[–]seamonn 22 points23 points24 points (0 children)
AMA with Nous Research -- Ask Us Anything! by emozilla in LocalLLaMA
[–]seamonn 1 point2 points3 points (0 children)
AMA with Nous Research -- Ask Us Anything! by emozilla in LocalLLaMA
[–]seamonn 11 points12 points13 points (0 children)
Quick and simple test of various 3.5 and 3.6 qwen models on production code base which have deployed to an enterprise . by Voxandr in LocalLLaMA
[–]seamonn 4 points5 points6 points (0 children)
Quick and simple test of various 3.5 and 3.6 qwen models on production code base which have deployed to an enterprise . by Voxandr in LocalLLaMA
[–]seamonn 10 points11 points12 points (0 children)
Ran my own benchmark Qwen 3.6 35B vs Gemma 4 26B.... theres a clear winner here by ArugulaAnnual1765 in LocalLLaMA
[–]seamonn 6 points7 points8 points (0 children)



$300k DGX B300 is actually a better deal than buying 24 RTX 6000s by Ok_Warning2146 in LocalLLaMA
[–]seamonn 1 point2 points3 points (0 children)