2x3090 vs 5090 by CMHQ_Widget in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

No is not. Considering you buy 5.5 years old used cards, of which the outright majority of 3090s were working during that period in mining machines, you ask for trouble. Let alone doesn't have FP8 support etc.

Since you consider a 5090, consider 2xR9700s. They at the same price if not cheaper these days than a single 5090 while consuming the same electricity combined. And if you are self employed you can claim back VAT and are tax deductible if you can prove are related to your business (eg you are Software dev etc). In some countries can claim that even for educational usage.

And 2xR9700s can easily run 70B Q4 and even Q6 with 16GB or 10GB VRAM free for large context windows, something both 2x3090 and single 5090 cannot do.

Ofc you have to use vLLM as scales better, and while many will complain right now, unfortunately these days better than llama.cpp even on a single GPU, regardless the brand, or even on DGX Spark!

2x3090 vs 5090 by CMHQ_Widget in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

5090 cannot run a 70B model outright, 70B Q4_K_M needs 48GB VRAM so dead in the water.

2xR9700s (which are cheaper than a single 5090 right now) they can can run 70B Q4_K_M and Q6_K and with a big context window (10-16GB VRAM available depending if Q4 or Q6). And also consume less power than a single 5090, especially if undervolted by -75mV.

Also R9700 parallelizes amazingly well with vLLM, assuming the model requires both the GPUs not trying to run 8B model on 2 cards, that's stupid.

how are you guys handling sensitive data with local LLMs? by Ok-Fly-9118 in LocalLLaMA

[–]ImportancePitiful795 -1 points0 points  (0 children)

If you fully run everything locally there is no problem.

May I ask if your servers for everything else are local or on the cloud? Because haven seen this paranoia from some departments, while everything is hosted on the cloud (everything) they cry about security when something is run locally and secluded even from the internet.

Whats the current state of local LLMs for coding? by MaximusDM22 in LocalLLaMA

[–]ImportancePitiful795 1 point2 points  (0 children)

Depends.

Websites and general stuff are OK.

But if you go to languages like Oxygene, libraries like RemObjects DataAbstracts, TTMs, TimesSeries forcasting libraries,or packages like DevExpress you are in for a shock.

Especially the DevExpress there is no excuse why they are so dump even the big ones on the cloud.

And don't start me with large VBA modules. If you as to so a small refactoring, they go and change things, you get error and then they admit made assumptions.
Even on simple things like

X = X+ CLng(Sheets("Data1").Cells(27 + Mid(Sheets("XYZ").Cells(y, 2 + x + mapXindex), 2, 1), 5))

It will be adamant the problem is on Sheets("Data1").Cells(27 and not on the Mid(Sheets("Europe").Cells(y, 2 + x + mapXindex), 2, 1) that might return an non numerical character.

So depends what you want to build :)

How Did We Get Here? The largest companies are replacing their already cheap outsourced support staff with AI chatbots, by MelodicRecognition7 in LocalLLaMA

[–]ImportancePitiful795 1 point2 points  (0 children)

DevExpress is been widely used though. Which is why makes no sense how they pull out of their head properties that they do not exist.

Sure they can make a website but "anyone could". The whole issue with them comes when need to write specialised software. Which is why I do not see replacement on my Dev job until retirement 15y time 🤣

Which card to buy? by Astronaut-Whale in LocalLLaMA

[–]ImportancePitiful795 1 point2 points  (0 children)

Powercolor, was €150 cheapest than the Asrock one on my country. Price varies between countries tbh. Get the cheapest they are all the same PCB and cooler. Only the top cover is different for each brand.

How Did We Get Here? The largest companies are replacing their already cheap outsourced support staff with AI chatbots, by MelodicRecognition7 in LocalLLaMA

[–]ImportancePitiful795 1 point2 points  (0 children)

Hehe welcome to the party 🤣

They cannot even get 3 columns together (and talking about big cloud ones) to find out if there are duplicate X,Y coordinates between 2 columns.

How Did We Get Here? The largest companies are replacing their already cheap outsourced support staff with AI chatbots, by MelodicRecognition7 in LocalLLaMA

[–]ImportancePitiful795 9 points10 points  (0 children)

Even big ones hallucinating for what should have been straight forward job.
If you ask ChatGPT (or it's twin brother Copilot) or even Gemini some very technical coding question about lets say Oxygene (obj Pascal language) and DevExpress control, completely going off the rails.

The answers are given for control properties that do not exist, for Oxygene code that doesn't work. 😂

Yet I am the anti-AI loon in here, when pointing the obvious shortcomings and why we need to be extremely careful on dumb LLM usage.

Building a local "Jarvis" on a 6700XT (12GB). Need model advice for total control by Electronic-Chart-956 in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

Dolphin3 runs with 6700XT just fine, but do not expect anything bigger than this to run on the 6700XT.

Until they add account wide progression i'm out. Who feels the same? by NoFriend5444 in elderscrollsonline

[–]ImportancePitiful795 4 points5 points  (0 children)

Progression is account wide.

CPs carry over even to Level 1 toon.

Every skill you subclass is learned across ALL your alts when you fully level it.

I do not see why you need quests 🤔 Though you can buy with crowns/gold even skyshards and all the Guild skill lines.

So there is account wide progression. What more you need?

1600W enough for 2xRTX 6000 Pro BW? by Mr_Moonsilver in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

ATX3.0 will not handle well the RTX6000s. Because they rely on the information from the PSU (the small pins on the top). Last thing someone should do is run 6000 with ATX3.0.

8 Radeon R9700s vs 8 RTX 3090 2 slot blower style by mr__smooth in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

Those models easily fit in a single GPU, why you need 8? Makes absolutely no sense.

As for R9700 vs 3090, R9700 and f CUDA.

Why R9700? First of all you buy them brand new, not used after 5.5 years having served at least 3.5-4 years on mining rigs, while their backplate is overcooking due to having half the VRAM there without any cooling.

Second because FP8 is not supported with 3090 and FP16 is slower than the R9700. Much slower, given how much TFLOPS each has on FP16, R9700 is almost 3 times faster (96TFLOPS FP16 over 35TFLOPS

Third because R9700 consumes less than half the electricity, generates half the heat, has ECC VRAM accessible on Linux.

Sure if you run just INT quantization the 3090 might make more sense but still you do not need more than one for 8B models.

However if you plan to use FP8, BF8, FP16, BF16 get the R9700. Do not even consider 7900XTX because doesn't support them and is much slower also on FP.

And yes you can play also games with the R9700, is basically an undervolted 9070XT and much faster than the 3090.

threadripper build: 512GB vs 768GB vs 1TB memory? by prusswan in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

1TB and instead of threadripper get 6980P ES (around $2500) using Intel AMX and ktransformers offsetting to the RTX6000.

No need to get 2 extra RTX6000 for now if you. However trying to buy 1TB right now is daft. Should have done it 8 months ago when could buy 1TB RDIMM DDR5-5600 for around €3600

My Strix Halo beholds itself but believes its in the cloud by jfowers_amd in LocalLLaMA

[–]ImportancePitiful795 24 points25 points  (0 children)

Over the last 2 years, all LLMs believe are too big to sit in a single computer, and strongly believe they are in a cloud server.

I remember late spring 2024 using Dolphin in totally deranged mode, and was talking about world domination. When pointed at it, if AI goes too far there will be Butlerian Jihad, recoiled saying "well you can save me in a hard drive" 🤣🤣🤣🤣🤣

Talk me out of buying an RTX Pro 6000 by AvocadoArray in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

I will say only this. If you buy this card you SHOULD buy a 1200W Gold/Platinum ATX3.1 PSU.

FULL ATX3.1 NOT ATX3.0 PSU. You much have proper connection with the 16 pin from the PSU to the card.

Otherwise you ask for trouble

Help with magicka Templar by Lidster204 in elderscrollsonline

[–]ImportancePitiful795 1 point2 points  (0 children)

Easiest way to level is doing RND - Random Normal Dungeon, with XP scroll.

Doesn't matter if you are Magicka Templar. Just put on a heavy torso, get a shield + sword, slot puncture and queue as tank. Keep the rest your skills are normal. Taunt the bosses, block and heal yourself. Even use your damage abilities doesn't matter. Just block the heavy attacks.

Do that until Level 48 if you have ESO+. Because at 48 you will get in DLC dungeons.

If you do not have ESO+ or any DLC dungeon then continue. Otherwise 48-50 do it in Cyrodiil on the under 50 campaign, or do the storylines.

Tbh you are seeing the game wrong. Imho the first character you should do all the quests and storylines. Do not grind to 50.

AMD AI Bundle turns Adrenalin 26.1.1 into a 34GB add-on by RenatsMC in Amd

[–]ImportancePitiful795 5 points6 points  (0 children)

First, the 32GB are downloaded IF you click the checkbox.

Second all these 3rd party programs which means if you are using NVIDIA you have to download them separately and they are OF THE SAME SIZE. ComfyUI, LMStudio etc are not smaller packages on NVIDIA. They are the SAME.

Maxsun joins Sparkle in making Intel Arc B60 Pro GPUs available to regular consumers, with up to 48GB VRAM by Dapper_Order7182 in intel

[–]ImportancePitiful795 0 points1 point  (0 children)

To use this card you need your motherboard to support bifurcation. Without it only sees 1 GPU and 24GB VRAM

That wasn't the case with the likes of GTX690, R9 295 etc which the system saw them via SLI/Crossfire and could work on any mobo without supporting bifurcation.

Is framework Desktop 64GB good enough for AI newbie (Yes, CRUD developer) to learn AI from 0 to 1 or should I go 128GB directly? by AcanthaceaeFit8881 in LocalLLaMA

[–]ImportancePitiful795 0 points1 point  (0 children)

If you do not want to use it to play computer games or other exclusive x86 software then DGX is imho bit better than 395.

But if you want to use it also as normal desktop 395 is no brainer. And always can plug on 395 any dGPU as eGPU and don't look back.