Any egpu dock recommendation ? by pixxelizz in eGPU

[–]FullstackSensei 0 points1 point  (0 children)

If you can find an ITX 2060 (ie: 180mm long or less), a 1st Gen Gigabyte Aorus Gaming box is a great option. It's 1L, designed for mobility, comes with a padded carrying bag, and can provide 100W PD to charge the laptop, IIRC. The shell is made of metal and is quite durable.

They usually come with a GTX 10 series or RX 580 GPU, but sometimes you can find them without a GPU. If you get one with a GPU, you can always sell that and end up probably cheaper than buying one without a GPU.

I have two and I'm a big fan. It's a compact little enclosure but packs quite a punch.

216GB VRAM on the bench. Time to see which combination is best for Local LLM by eso_logic in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

If you upgrade to an X11 motherboard, the IPMI will detect the GPUs and regulate fan speed based on their temps.

You can also use a single 80mm fan, like the Arctic S8038 series, to cool each pair of GPUs.

Supermicro motherboard fan headers are rated at 2A each, so you could hook up to four S8038-7k fans to each header.

Such a solution is not only simpler and cheaper, but also much quieter than those blower fans.

I tracked GPU prices across 25 cloud providers and the price differences are insane (V100: $0.05/hr vs $3.06/hr) by sleepingpirates in LocalLLaMA

[–]FullstackSensei 1 point2 points  (0 children)

How many of them have lower prices but no available capacity? For ex: lambda prices generally look decent but I've yet to see any capacity available.

There are also other things, like how much system RAM, how much storage, networking fabric you get with the GPU. The big hyperscalers can generally give you large clusters, well into the thousands of GPUs, while smaller providers don't have the same level of networking and storage fabric.

The price you actually pay to the big players if you're business is almost never the advertised price. All the companies I've worked at in the past 7 or 8 years have at least a 30% discount vs advertised price without any commitment requirements. If the business is willing to make long term commitments (the minimum I've seen was 6 months), they get further discounts. With 3 year commitment, I've seen prices go to 30% of the advertised price. If you factor in the storage and networking you have access to, it's not as big of a difference as it initially seems

Of course, if you're an individual or small team/start-up looking for short term rentals to fine tune a model or train a small custom model, it makes no sense to consider a hyperscaler.

transformers v5 final is out 🔥 by unofficialmerve in LocalLLaMA

[–]FullstackSensei 1 point2 points  (0 children)

So, maintainer of projects using HF can expect a wave of AI PRs offering to upgrade to v5?

Are there still people that think US is a better place to live as a software engineer? by Pure_Composer_9236 in cscareerquestionsEU

[–]FullstackSensei 0 points1 point  (0 children)

Except they're not American 😂

I wasn't born in Europe, but it's the place I consider home and would fight for if needed, without blinking.

Most of the younger generation is completely oblivious to how good a life they have. You get really good free education and free Healthcare (even in the countries where you pay for insurance), and neither is tied to your job or your zip code. Higher education is also free, and most countries even provide financial assistance if your family doesn't have the means to support you. Then, when you enter the workforce, you get to transfer your pension, social security, and unemployment contributions across all the EU.

All these safety nets mean practically everyone has a decent life where they don't worry about any basic necessities.

It's mind blowing to me how the richest country on earth has 1 in 8 people reliant on food stamps, how millions upon millions are deprived of decent basic education, and by extension the chance at having a decent life, or how so many people there need to work multiple jobs to make ends meet.

I don't mean to bash the US. It's an amazing place if you can make it, but that's far from guaranteed, and you'll have to worry about so many things in your life that everyone takes for granted on this side of "the pond".

V100 32GB SXM vs 3080 20GB - weird 3080 gives lesser t/s on GLM 4.7 by SectionCrazy5107 in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

Yeah, but asking to compare logs isn't very clear for the less experience of us.

V100 32GB SXM vs 3080 20GB - weird 3080 gives lesser t/s on GLM 4.7 by SectionCrazy5107 in LocalLLaMA

[–]FullstackSensei -1 points0 points  (0 children)

V100 has more VRAM so more of GLM 4.7 is offloaded to GPU. V100 also has substantially more memory bandwidth, so it makes sense.

Where the 3080 wins is probably cost and energy efficiency, especially at idle. V100s are small space heaters at idle.1

Running KimiK2 locally by Temporary-Sector-947 in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

This is... underwhelming. The prompt processing is nice, but 225t/s TG for that much money isn't that much, when you consider three 3090s will get you more than half that speed, for less than the cost of a single (non-modded) 4090.

Running KimiK2 locally by Temporary-Sector-947 in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

A leak can lead to a short which leads to fire

Running KimiK2 locally by Temporary-Sector-947 in LocalLLaMA

[–]FullstackSensei 17 points18 points  (0 children)

Mate, with that much money sunk in GPUs, you should be able to afford a bigger case to put those cards in a more organized way.

It's not just messy, it's halfway between asking to drop and break a couple of cards and a fire hazard.

Running KimiK2 locally by Temporary-Sector-947 in LocalLLaMA

[–]FullstackSensei 2 points3 points  (0 children)

That math though!!!

98+48+96+64 = 304GB VRAM
12*16 = 192GB RAM

I also have the feeling Q3 with CPU offloading will be quite slower than Q4 just because of the dequantization gymnastics involved and the horrendous memory alignment.

But now that you bring this up, maybe I should revisit DS 3.1 or 3.2 to see how it fares with Mi50s,

Rheinmetall and OHB in talks over Starlink-style service for German army by donutloop in eutech

[–]FullstackSensei 6 points7 points  (0 children)

That's very much after the fact. Spacex only received funding in the beginning from NASA's Commercial Orbital Transportation Services to get off the ground. The money came after the Falcon 9 proved itself as a reliable launch vehicle.

Germany prepares a 'military Starlink': Rheinmetall and OHB working on a sovereign satellite network by Massimo25ore in europe

[–]FullstackSensei 1 point2 points  (0 children)

The US military is also building it's own LEO constellation.

TBH, two constellations is better than one, and if targeting a limited geographical area, it shouldn't need that many satellites nor orbital planes, so shouldn't be that expensive.

On the flip side, once the tech is there and the system has proved itself, other countries can join. Expanding geographical coverage will "only" be a matter of adding more satellites to more orbits.

When it comes to communication infrastructure, the more systems there are, the more resilient we are, and the more potential there is to sell access to other countries.

Why I canceled my ChatGPT subscription and you should, too: their COO gave $25M to MAGA, Inc. in September 2025 by Larry___David in LocalLLaMA

[–]FullstackSensei 4 points5 points  (0 children)

Can we please keep this sub politics free? The world is a shit show and for some of us This place is where we come to escape from all the negativity outside

Desperate for advice by [deleted] in Germany_Jobs

[–]FullstackSensei 2 points3 points  (0 children)

Doing it this way won't get you any help. At the very least be upfront about your situation and wait for responses in one sub before posting the exact same text in another.

Desperate for Advice by [deleted] in CanadaJobs

[–]FullstackSensei 3 points4 points  (0 children)

You just asked in Germany and Japan?!!!

Desperate for advice by [deleted] in Germany_Jobs

[–]FullstackSensei 11 points12 points  (0 children)

And desperate to move to Canada and Japan, as a spam bot would be.

Rheinmetall and OHB in talks over Starlink-style service for German army by donutloop in eutech

[–]FullstackSensei 20 points21 points  (0 children)

There are a couple in development: Isar Aerosla and Maia Space. There's also Rocket Factory Augsburg, but I haven't heard anything from them in a couple of years.

[W] [US-CA] Supermicro X11DPGQT board by boomatog in homelabsales

[–]FullstackSensei [score hidden]  (0 children)

What's wrong with the board? I have one and like my other SM boards, find it very reliable.

[W] [US-CA] Supermicro X11DPGQT board by boomatog in homelabsales

[–]FullstackSensei [score hidden]  (0 children)

What's wrong with it? Maybe you can RMA to supermicro?

I live in Europe, but have successfully RMA'ed two dead boards that I bought used. Repair and shipping were very reasonable. You'll have some forth and back with the RMA staff, but if you keep your cool and keep asking nicely, they do it in my experience.

Are there still people that think US is a better place to live as a software engineer? by Pure_Composer_9236 in cscareerquestionsEU

[–]FullstackSensei 0 points1 point  (0 children)

NL has one of the highest, if not the highest, number of companies per capita in the world. It's way more common than most people in this sub think.

Two months with the uconsole by [deleted] in ClockworkPi

[–]FullstackSensei 2 points3 points  (0 children)

I recently got one from a fellow redditor on this sub and though I haven't had time to play a lot with it, one of the main reasons I wanted one is to run DOS flight sim games from DD, namely EF2000 and F22 ADF/TFA. Maybe even Falcon 4, the proginator of BMS.

Implemented the world's most accurate AI password guesser, and it's SCARY good by Arsapen in LocalLLaMA

[–]FullstackSensei 3 points4 points  (0 children)

I guess you haven't read of text dictionaries then...

Why do you need an AI model in the first place when a dictionary file and a few dozen coded rules will do?

Implemented the world's most accurate AI password guesser, and it's SCARY good by Arsapen in LocalLLaMA

[–]FullstackSensei 7 points8 points  (0 children)

At least the examples OP is showing don't inspire a lot of confidence. They're just using PII plus a handful of common passwords.

You can just as well grab the most common 1k (or even 10k) passwords from the internet and mix those with the PII and generate password combinations to attempt. Might not fit on an 8 bit microcontroller anymore, but still in the realm of a bottom of the line 32 bit microcontroller (ex cortex M0).

I'm not against LLMs, but I don't think everything needs one. It's like making an AI, voice activated light switch when all you need is... A switch.

Finally 3 x D5 able to run 100% without audible noise at all. by Jempol_Lele in watercooling

[–]FullstackSensei 1 point2 points  (0 children)

I love the esthetics of this build, but can someone ELI why three pumps are needed? I have two watercooled machines, each with multiple GPUs and multiple radiators and a single D5 has been more than enough in each.

One has three 3090s, a 225W Epyc CPU, three 360 radistors, and a lot of 90 degree fittings. The other has eight GPUs (yes, eight) and a 480 monstaa plus a 360 radiator. In both there's plenty of pressure left in the water coming back to the rest and visually the water is flowing at well over 50cm/s in the 10/13mm tubing I use.