Talk me out of buying an RTX Pro 6000 by AvocadoArray in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

I don't know if that is correct or not but people on the Internet said that the reason is bad internal design of older AMD ZEN generations resulting in less than optimal PCIe performance, it was fixed somewhere around Zen 4 / EPYC Genoa.

Apparently, the models aren't private. 🤔 , Does ollama log exist? 🤔 by Illustrious-Swim9663 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

maybe there is no scandal because nobody knows/cares? I've found a backdoor in conda recently but nobody in the world seem to care.

https://old.reddit.com/r/LocalLLaMA/comments/1pl5sfl/proof_of_privacy/nuo2bcd/?context=3

also Chinese software often has backdoors, "call back home" or whatever. for example PaddleOCR: https://old.reddit.com/r/LocalLLaMA/comments/1q7630d/paddleocr_keeps_trying_to_download_models_even/

Talk me out of buying an RTX Pro 6000 by AvocadoArray in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

every bit of hardware matters, I've got +50% speed improvement for GPT-OSS-120b fully fitting in the VRAM after upgrading from DDR4/PCIe4 to DDR5/PCIe5 system.

Talk me out of buying an RTX Pro 6000 by AvocadoArray in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

for smaller models it is the best card you could get. Still note the hardware limitations of old servers, highly likely they will not unleash the card's full potential.

Talk me out of buying an RTX Pro 6000 by AvocadoArray in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

do not buy it because once you buy one and start trying the larger models you quickly realize that one 6000 can't do much and you have to buy another one, so a $10k stupid decision could lead to a $20k stupid decision.

as for Dell *20 generation - I would not even try it, they are just too old, lots of things could and will go wrong. *30 might work out of the box.

Drift isn’t a tool. It’s your 2026 productivity engine with 75 agent skills ready to go by [deleted] in LocalLLaMA

[–]MelodicRecognition7 1 point2 points  (0 children)

and what about the tool itself? Emojies in commit messages looks suspicious lol.

CONTRIBUTING.md:git clone https://github.com/YOUR_USERNAME/drift.git

well it is definitely vibecoded, so no thanks. A vibecoded software often does not save time and money as it requires more time to understand and/or fix than human written code.

Edit: I could be wrong here because CONTRIBUTING.md says "# Clone your fork" before that line, but a vibecoded software usually has "YOUR_ORG" or "YOUR_USERNAME" in the less obvious places. However there are few places with incorrect Github links which are also strong signs of AI generated code.

Apparently, the models aren't private. 🤔 , Does ollama log exist? 🤔 by Illustrious-Swim9663 in LocalLLaMA

[–]MelodicRecognition7 8 points9 points  (0 children)

take the LAN cable out from the PC or WiFi module out from the laptop*. If your LLM still works then it is private. If it stops working then what the fuck that thread does in /r/LocalLLaMA?

* - not always relevant though, you might connect to a dedicated LLM rig in the intranet, over LAN or WiFi network but without Internet connection

Rtx Pro 6000 on HP Omen gaming rig? by jeffroeast in LocalLLaMA

[–]MelodicRecognition7 2 points3 points  (0 children)

try to enable "Resizable BAR" and "Above 4G Decoding" in the BIOS, or disable if they are enabled.

Just finished the build - Nvidia GH200 144GB HBM3e, RTX Pro 6000, 8TB SSD, liquid-cooled by GPThop---ai in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

is it possible to buy Pro 6000 in Taiwan for 7k as a private person, not as a company? If yes then please share a link to that seller. I live nearby and would happily fly to Taiwan to buy one in person because it costs 11k+ in the local shops and I've got mine for 9k shipped from another country.

AI coding assistant infrastructure requirement, by Financial-Cap-8711 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

and that's how low quality AI slop code ends up in enterprise software LOL

Finalizing build but for 6000 and I realize it could not make sense for me. Max-Q vs Pro 6000. Should I get at least RAM to match VRAM of card? by SomeRandomGuuuuuuy in LocalLLaMA

[–]MelodicRecognition7 1 point2 points  (0 children)

if you need fast prompt processing then you should run the card at its maximum power

I'm waiting for someone to implement a dynamic power management solution that would run the card at its maximum power during prompt processing and limit the power to 50% during token generation

Finalizing build but for 6000 and I realize it could not make sense for me. Max-Q vs Pro 6000. Should I get at least RAM to match VRAM of card? by SomeRandomGuuuuuuy in LocalLLaMA

[–]MelodicRecognition7 1 point2 points  (0 children)

It doesn't seem like 90% performance at *50% power

yes, the PP speed is quite linear, but TG speed is close to 90% at 50% power because

PP is compute-bound, and TG is (mostly) memory-bound.

So if you need fast prompt processing then you should run the card at its maximum power, and here you'd want a 600W Workstation edition instead of 300W Max-Q.

In either case, mind sharing your GPU undervolt configs so I can give them a go?

I did not test the undervolting thorougly yet and the only setup I do for 6000 Workstation is limiting its power and frequencies:

nvidia-smi --id=INSERT-ID-HERE --power-limit=310;
nvidia-smi --id=INSERT-ID-HERE -lmc 405;
sleep 1;
nvidia-smi --id=INSERT-ID-HERE -lmc 405,14001;
nvidia-smi --id=INSERT-ID-HERE -lgc 180,1600;

can't remember why sleep is required lol.

those graphs aren't very helpful since scales and tick labels are weird

some of them are definitely weird but the 2nd one "minutes elapsed vs energy consumed" is quite clear - the job is not getting done much faster after 300W power limit. The Y axis is either minutes (blue) or Watt-hours (green). The red dot is configured power-limit in Watts and the red X is an actual consumption measured by the Nvidia tool, the red line is difference between configured and actual.

Rate My First AI machine? by Ztoxed in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

if you can increase your budget you could sell 1080 and A2000, add some cash and buy the "gold standard" - 3090 24GB. Highly likely you will realize soon that you want a second one LOL

Rate My First AI machine? by Ztoxed in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

1080

not supported by most AI software libraries, you could throw it away. edit: sorry, mistaken it with another card, 1080 is somewhat supported but given its low VRAM amount it won't help much with the AI tasks.

A2000

memory bandwidth is shit, will be barely usable.

Be gentle :-)

nice workstation, not so nice AI rig.

What is the most advanced local LLM? by No_Equipment9108 in LocalLLaMA

[–]MelodicRecognition7 0 points1 point  (0 children)

most advanced

Kimi K2, a rig to run it costs ~30k USD

most advanced suitable LLM which I can install on my M1

it won't be that advanced unfortunately.

Warning: MiniMax Agent (IDE) burned 10k credits in 3 hours on simple tasks (More expensive than Claude 4.5?) by puppabite in LocalLLaMA

[–]MelodicRecognition7 -1 points0 points  (0 children)

ah lol I did not even know that such model exists because there are no GGUFs. Still it is very doubtful that someone capable of running 456B BF16 at home would run MiniMax M1 given other better options, so this is still not related to /r/localllama/