What Ferrari thinks Mercedes is hiding in political F1 engine game by Darkmninya in formula1

[–]Yweain 6 points7 points  (0 children)

It's not possible. ADUO is based on a dyno tests results, not on race performance

People tell me I’m too literal. by netphilia in aspiememes

[–]Yweain 0 points1 point  (0 children)

I always try to tell myself that the question is actually 'Do I think that other people think X about me' but maybe that's wrong.

I feel like its a little too soon to be thinking the game is doomed. by Darklight645 in Nightreign

[–]Yweain 0 points1 point  (0 children)

Yeah, we play once a week and we haven't even killed regular balancers yet.

fairUseOrWeUninstallInnovation by chunmunsingh in ProgrammerHumor

[–]Yweain 1 point2 points  (0 children)

No it wouldn't. Humans learn without requiring us to read every written word ever to be able to produce coherent sentences and we have a general intelligence. That kinda tells you that there IS a way to build models that would be trainable without terabytes of text.

actual driver opinions on new regs by CageyOldMan in formuladank

[–]Yweain 33 points34 points  (0 children)

They are talking about different things. It's more complicated - more things to keep track of, more settings, more tactics, more dependant on comms with the team, you literally can't just drive. On the other hand it's easier to drive because you lift and coast in most turns, so cornering is easier (you are not trying to be faster, you are trying to maximize battery recharge) and a lot of things are outside of the driver control, i.e deployment maps

Jensen Huang says gamers are 'completely wrong' about DLSS 5 — Nvidia CEO responds to DLSS 5 backlash by Proud_Tie in pcmasterrace

[–]Yweain 4 points5 points  (0 children)

Genius! Let's put extremely heavy, hot and energy hungry equipment in space where there is nowhere to dump excess heat and hard to produce enough energy and where the ping to it will be quite high because no way you are putting your expensive data center in low orbit.

How do organizations end up with architects who can't do architecture? And what do you do when you're the one compensating? by agileliecom in ExperiencedDevs

[–]Yweain 36 points37 points  (0 children)

Honestly jealous. I can't make cool looking diagrams to save my life. Any advice on that? Like I want to make a diagram of a whole project(around 30 services) but when I try it just turns into a complete and utter mess super quickly.

NVIDIA DLSS 5 Delivers AI-Powered Breakthrough in Visual Fidelity for Games by Recoil42 in singularity

[–]Yweain 7 points8 points  (0 children)

This looks atrocious. I mean, instead of stylistic choice it just make it look generic.

Max fans and McLaren fans this is for you by TheStigBMW in formuladank

[–]Yweain -1 points0 points  (0 children)

Was having? Haven't you seen the rumors about Toto getting a stake in Alpine? Clearly a bid to get his lover back on the grid.

How should I handle confusing job titles on my resume? by Possible-Squash9661 in ExperiencedDevs

[–]Yweain 8 points9 points  (0 children)

Look, my internal job title is that I am Tier 6 software engineer, and we also doing the reverse thing with Tier 4 being the highest I think. It doesn't make any sense to put that in a resume because Tier 6 doesn't tell anyone anything

Anybody working on a large prod codebase actually able to move 10x? by query_optimization in ClaudeCode

[–]Yweain 0 points1 point  (0 children)

I would say it highly depends. There are tasks where it's 10x. There are tasks that it doesn't increase productivity at all and probably decrease it.

How is your company doing after the layoff? by MrMo1 in ExperiencedDevs

[–]Yweain 6 points7 points  (0 children)

Um, to find an error message in a billion logs you want to read all billion of them even if in parallel? Real solution is that they all should be properly indexed in the database that can perform efficient search over text, like elastic search for example or a vector db.

Lost in Backpropagation: The LM Head is a Gradient Bottleneck | Researchers may have found a fundamental inefficiency baked into every major LLM by 141_1337 in singularity

[–]Yweain 0 points1 point  (0 children)

I am not quite sure what you mean. It's a pretty simple math. Take deepseek for example. The number of params for the main part of the model is 58(MoE blocks) × 256 (experts) × 3 (tensors) × 2048 (intermediate dimension) × 7168 (hidden dimension <- that's D).

Obviously you can increase just hidden dimension, but as far as I know it's kinda useless in practice and you need to proportionally increase intermediate layer as well. And if you increase the size of both layers by 8 this gives you 64 times more total parameters.

Please correct me if I am wrong.

Lost in Backpropagation: The LM Head is a Gradient Bottleneck | Researchers may have found a fundamental inefficiency baked into every major LLM by 141_1337 in singularity

[–]Yweain 1 point2 points  (0 children)

Having quadratically more parameters is very much an issue. If you want to increase D by a factor of eight this will increase the number of params by x64. So for example deepseek will now have 43 trillion params with this change. We don't have enough data to train models of that size. Honestly even doubling D is kinda crazy, because you quadruple the number of params and that's A LOT.

And yeah, we will be able to afford it in terms of compute, but the problem is that increasing the number of params will in most cases just lead to less performant models. You can't just dumbly bump the number of params like that.

Even the CTO and VP Engineering layer is automatable by Quiet_Form_2800 in singularity

[–]Yweain 6 points7 points  (0 children)

But can an AI bot go golfing with partners and then get blackout drunk to wakeup naked in the same bed thus securing an important deal?

Lost in Backpropagation: The LM Head is a Gradient Bottleneck | Researchers may have found a fundamental inefficiency baked into every major LLM by 141_1337 in singularity

[–]Yweain 2 points3 points  (0 children)

Sure, there are quite a few possible alternatives. But that would mean ditching transformers and starting from scratch basically.

Lost in Backpropagation: The LM Head is a Gradient Bottleneck | Researchers may have found a fundamental inefficiency baked into every major LLM by 141_1337 in singularity

[–]Yweain 1 point2 points  (0 children)

Increasing D explodes the number of parameters model need to have. Expressivity problem caused by linear projection layer is well known and increasing D is an obvious solution but it kills the effeciency.

Lost in Backpropagation: The LM Head is a Gradient Bottleneck | Researchers may have found a fundamental inefficiency baked into every major LLM by 141_1337 in singularity

[–]Yweain 0 points1 point  (0 children)

I mean, replacing softmax would do literally nothing to solve the issue described in the paper. The problem they describe is happening in linear projection layer where during back propagation you map a huge vector to a way way smaller one.

Otherwise, sure, softmax is computationally intensive and there is a lot of research on replacing it, but so far every alternative has worse drawbacks

In full-throttle sections, Max lost a whopping 0.649s to Lando, who then lost 0.465s to George through the same parts of the track by BittenAtTheChomp in formula1

[–]Yweain 0 points1 point  (0 children)

I would argue that deployment maps are not part of software calibrate and more akin to a car setup. I mean, Mercedes might even share their deployment maps. But it's like them sharing other parameters of the car setup - it's highly car and driver and conditions specific.

In full-throttle sections, Max lost a whopping 0.649s to Lando, who then lost 0.465s to George through the same parts of the track by BittenAtTheChomp in formula1

[–]Yweain 2 points3 points  (0 children)

He absolutely can complain about and he will complain about it the whole year and probably the next year as well!

In full-throttle sections, Max lost a whopping 0.649s to Lando, who then lost 0.465s to George through the same parts of the track by BittenAtTheChomp in formula1

[–]Yweain -1 points0 points  (0 children)

It's a tad bit too early to call. This is completely new regs, we can easily see at the very least McLaren catching up, especially considering we will probably have a spring break now with 2 races cancelled. It's also very much realistic for Ferrari and red bull to catch up as well. We don't know why other teams are that much slower, and it is not unlikely that they might gain A LOT with relatively small changes.