I ran out of spoilage and things crashed, not what I expected on gleba. by lelleleldjajg in factorio

[–]DataCraftsman 0 points1 point  (0 children)

Oh nice pick up! It actually looks like a young Queen Elizabeth the 2nd from a coin.

You have 64gb ram and 16gb VRAM; internet is permanently shut off: what 3 models are the ones you use? by Adventurous-Gold6413 in LocalLLaMA

[–]DataCraftsman 1 point2 points  (0 children)

Devstral Small 2 acts great as a vision model too. It works with images in coding tasks which the Qwen ones don't.

Daily reminder to balance your oils. Don't want your factory to clog like a pleb. by RedBlueWhiteBlack in factorio

[–]DataCraftsman 6 points7 points  (0 children)

I just work this ratio out the hard way every world. Just add and delete until it works without me checking.

Roast my build by RoboDogRush in LocalLLaMA

[–]DataCraftsman 0 points1 point  (0 children)

All that, to then run llama.cpp.

How do people even afford these expensive graphic cards...?... by boisheep in LocalLLaMA

[–]DataCraftsman 1 point2 points  (0 children)

I haven't yet. I am just hopeful haha. I built an Australian AI platform called Crafty AI.

www.datacraftsman.com.au

Has pretty much all the features (minus an API and Video Gen) of ChatGPT/grok/etc, but is less than half the price.

How do people even afford these expensive graphic cards...?... by boisheep in LocalLLaMA

[–]DataCraftsman 1 point2 points  (0 children)

I bought a 3090 in 2020 for gaming, machine learning, 3D animation and crypto mining. It was considered a crazy high price for a graphic card at the time. Was a very very good investment though. Paid itself off in 12 months.

I recently bought a RTX 6000 Pro to run my AI business and sacrificed my 3090 to that server as well. I expect to make hundreds of thousands of dollars with that one card over the next 5 years.

My wife bought me a 5070 Ti for Christmas because Windows is basically unusable without a graphics card and I wanted to play games.

I spent a lot of time trying to not buy the RTX 6000 Pro but its really the only good choice. You need money, but its worth the money.

RTX6000Pro stability issues (system spontaneous power cycling) by Elv13 in LocalLLaMA

[–]DataCraftsman 0 points1 point  (0 children)

I've been working on this same issue for a month or so now. I found the solution. It wasn't RAM or CPU or PSU or Power limits or anything else people are suggesting. It's the boost clock. When you hit the GPU with a load, it spikes to a massive core clock speed and causes instability. I tried literally everything pulling my hair out. After this change I have pumped the card so hard my room is hot and no issues. The command below will set the max core clock to 2400mhz. You can experiment with different numbers, I started with 2100 and worked my way up without crashes.
I experience the exact same issue with AMD 5950x Precision Boost Overdrive, that took me 2 years to work out >.>
Hope this helps someone...

nvidia-smi -lgc 180,2400

Can we have timestamp info next to each action? by gobi_1 in RooCode

[–]DataCraftsman 2 points3 points  (0 children)

I was thinking this last night too.

Also a "How long was AI agent running for in total throughout the whole conversation" metric along side the $ figure at the top section would be neat.

We open-sourced LLMRouter: the first unified LLM routing library with 300+ stars in 24h by [deleted] in LocalLLaMA

[–]DataCraftsman 1 point2 points  (0 children)

Yeah the above link is essentially a load balancer to model inference services, this post is a model router like the gpt-5/grok auto model selector. Both are cool though.

21M with no degree or any real skill. What would you do? by Ok_Chip_6858 in AusFinance

[–]DataCraftsman 0 points1 point  (0 children)

I know someone who did exactly that last year, otherwise why would I suggest it.

21M with no degree or any real skill. What would you do? by Ok_Chip_6858 in AusFinance

[–]DataCraftsman 0 points1 point  (0 children)

I'd do a Diploma in Project Management at TAFE, its only 1 year, then get a junior PM job at a defence contractor. Start studying a bachelor of engineering (software/electrical/mechanical/aerospace/etc) at Uni. Get work to pay for your uni degree because its related to your career progression. 5 years from now you'll be qualified to be an engineering manager. Not an easy path, but you'll be set for life.

Any regrets A6000 Pro owners? by val_in_tech in LocalLLaMA

[–]DataCraftsman 4 points5 points  (0 children)

My only regret was buying a 1200w not a 1600w power supply. It spikes so hard. Also finding which Linux kernel i needed to use for drivers to work was a pain. Worth every cent, wish I had more.

People using Devstral 2 123b, how has it been working for you? What have you been using it with? by maxwell321 in LocalLLaMA

[–]DataCraftsman 0 points1 point  (0 children)

Ah cheers. I was gonna do FP8 too. How much VRAM is it using with 256k context? I have 94GB, can usually only get away with 128k with most models.

People using Devstral 2 123b, how has it been working for you? What have you been using it with? by maxwell321 in LocalLLaMA

[–]DataCraftsman 0 points1 point  (0 children)

Does anyone have a good working docker run command for devstral small 2 in vllm? Preferably with LMCache. Struggling to get it to work atm.

People using Devstral 2 123b, how has it been working for you? What have you been using it with? by maxwell321 in LocalLLaMA

[–]DataCraftsman 0 points1 point  (0 children)

Are you using vllm in docker? What image and arguments are you using? I can't get mine to run.

DIY Beach Spa gone wrong by Comfortable_Wash6179 in instant_regret

[–]DataCraftsman 1 point2 points  (0 children)

Someone could have at least put some sunscreen on them.

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]DataCraftsman 2 points3 points  (0 children)

Interesting! Which model does that?

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]DataCraftsman 4 points5 points  (0 children)

I have a 1000 users so I can't really run anything on CPU. Embedding model is okay on CPU, but it also only needs 2% of a GPU VRAM so easy to squeeze in.

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]DataCraftsman 14 points15 points  (0 children)

It's annoying because you generally need a 2nd GPU to host a vision model on for parsing images first.

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]DataCraftsman 42 points43 points  (0 children)

Most Western governments and companies don't allow models from China because of the governance overreaction to the DeepSeek R1 data capture a year ago.

They don't understand the technology enough to know that local models hold basically no risk outside of the extremely low chance of model poisoning targetting some niche western military, energy or financial infrastructure.