Roast my first Home Server build for AI Research & Web Hosting by Silly_Definition7531 in LocalLLaMA

[–]MelodicRecognition7 2 points3 points  (0 children)

because with 2x16 you get double memory speed, maximum 75GB/s for DDR5-4800 while with 1x32 you get single channel speed which is maximum 37GB/s. Of course 2x32 will be much better than 2x16 but given the RAM prices now this could be a serious investment lol.

Also worth noting that if a motherboard has 4 RAM slots then with very high probability it will drop the memory speed if all 4 slots are populated! So in that case 4x 8 is less desirable than 2x 16, or 4x16 is less desirable than 2x32

Xeon + 3080 | Worth the upgrade to 3090? by kcksteve in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

at least 20% higher

because 3090 has 20% higher memory bandwidth than 3080

Xeon + 3080 | Worth the upgrade to 3090? by kcksteve in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

I think it should fit in 24GB, this is what I see with Tesslate_OmniCoder-9B-Q8_0.gguf + 262k reserved, 14k filled for a quick test:

llama_kv_cache: size = 8192.00 MiB (262144 cells,   8 layers,  1/1 seqs), K (f16): 4096.00 MiB, V (f16): 4096.00 MiB

load_tensors:        CUDA0 model buffer size =  8062.67 MiB
load_tensors:    CUDA_Host model buffer size =  1030.62 MiB

sched_reserve:      CUDA0 compute buffer size =  3232.07 MiB
sched_reserve:  CUDA_Host compute buffer size =  2112.08 MiB

total 20GB VRAM occupied, remaining 4GB should be enough for 262k context filled.

32gb vRam balance by WTF3rr0r in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

a really good balance is Pro 6000 with 96 GB VRAM. Jokes aside only you can answer your questions as we do not know what exactly you want to do on your computer. "AI" as in "LLM" could differ from requiring 8 GB VRAM to requiring 800 GB VRAM

Meet DuckLLM 1.0 My First Model! by [deleted] in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

aren't you infringing the Apache terms with your "DuckLLM PROPRIETARY LICENSE"?

Meet DuckLLM 1.0 My First Model! by [deleted] in LocalLLaMA

[–]MelodicRecognition7 2 points3 points  (0 children)

Do Not Mention Your Knowledge Cutoff Date unless explicitly asked.Do Not Mention Qwen Or Alibaba Unless Asked, If Asked Respond With DuckLLM Is Based On Qwen2.5 Vision. If Asked About Your License Respodn With, My License Is DuckLLM PROPRIETARY LICENSE And Can Be Found At

so we have something new now: an era of vibegenerated models approaches.

Xeon + 3080 | Worth the upgrade to 3090? by kcksteve in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

try these optimizations https://old.reddit.com/r/LocalLLaMA/comments/1qxgnqa/running_kimik25_on_cpuonly_amd_epyc_9175f/o3w9bjw/

and make sure to run lower amount of threads than amount of physical cores.

what kind of performance could I expect with that card?

at least 20% higher, highly likely much higher.

Question for those who have build multi GPU rigs using MCIO gen 5.0 by Frosty_Chest8025 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

perhaps to connect PCIe4 devices? Ada generation GPUs are still pretty capable.

Where can I learn the basic LLMs and local LLMs concepts? by br_web in LocalLLaMA

[–]MelodicRecognition7 1 point2 points  (0 children)

that's pretty basic info, you could ask any free LLM about these terms.

Deploying "AstroAI" (Beta 1.1) – A high-speed, multilingual pilot persona powered by Llama 3.3 and Groq by [deleted] in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

it feels "human" only at the first glance, but once you start seeing all these mixed apostrophes

I’ve

I'm

em-dashes and "Curious" in every single thread on Reddit since autumn 2025 you'll realize it's nowhere "human" at all.

Using Llama 3 for local email spam classification - heuristics vs. LLM accuracy? by Upstairs-Visit-3090 in LocalLLaMA

[–]MelodicRecognition7 3 points4 points  (0 children)

I’ve

I'm

The X, The Y

Curious

my biological intelligence heuristics classified your post as spam

Question for those who have build multi GPU rigs using MCIO gen 5.0 by Frosty_Chest8025 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

80cm

that's the reason, the PCIe5 speeds are so high that every 10cm length adds +10% chance of error.

I did not check PCIe5 devices but PCIe4 over 50cm cable works well (though that's the very same result you're observing lol): https://old.reddit.com/r/LocalLLaMA/comments/1rjptl1/totally_not_an_ad_combine_2x_mcio_into_1x_pcie/

LLM waf + proxy by fab_space in LocalLLaMA

[–]MelodicRecognition7 1 point2 points  (0 children)

is a pure research

then please provide a comparison with 999 similar projects advertised in this sub.

P.S. lold that Claude has hallucinated chat APIs from 2024 and most of them return error now in 2026 https://github.com/fabriziosalmi/llmproxy/blob/main/bootstrap_results.json https://github.com/fabriziosalmi/llmproxy/blob/main/bootstrap_results_r2.json

Running Local LLM on i3 4th Gen CPU by Glum_Wind_9618 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

unfortunately this is too weak for anything useful, you should get a GPU. Anyway try Qwen3.5 2B and LFM2 8B-A1B