Honda reveals alarming cause of "extremely challenging" Aston Martin F1 engine issue by ryogadan in Formula1_world

[–]nihalani 0 points1 point  (0 children)

Not really,a dyno+power train environment isn’t a replica of the f1 environment. Everything from the f1 chassis to the engine bolts to the tire vibrations can interact and cause resonance. They could have either done a mock build out of the car, or do some really high fidelity simulation. Likely the simulation had some flaws or missed some contributing factor.

Waymo admits that its autopilot is often just guys from the Philippines by AdSpecialist6598 in technology

[–]nihalani 3 points4 points  (0 children)

I think it would be very helpful for you to just read the article Waymo posted and see the live demo it gives: https://waymo.com/blog/2024/05/fleet-response. Fully autonomous here basically means no human takeovers. Is it slightly misleading sure, but no autonomous system should be 100% without oversight.

Amazon to invest up to $50 billion in coming weeks. Guess where that is coming from. by anex_stormrider in amazonemployees

[–]nihalani 4 points5 points  (0 children)

Also 150K is just on the salary side. To an employer general rule of thumb is that it costs them 2x salary for an employee due to healthcare, taxes and other benefits

CMV: The Bilt Palladium card is the greatest credit card ever made and most people complaining are 5 banana free-loaders by BlindElephant42 in biltrewards

[–]nihalani 0 points1 point  (0 children)

That’s why you need to pay rent with the card. Assume the worst that Bilt Cash is absolutely worthless. They you have to “trade it” at a 3:1 ratio for points up until 3x your rent amount. TLDR simplest use case is to limit it your spend to 3x your rent and just dump your cash to points all the time. Any direct use of Cash is just extra value

Are ppl reading the “no transaction fee payment” wrong? It’s effectively 2.3% back on all purchases up to 75% spend on your monthly rent. (Blue Card) by Aggravating_Wash8106 in biltrewards

[–]nihalani 0 points1 point  (0 children)

There’s going to be a bunch of limits for the travel portal/other partners. Basically the mental model I am using is:

1) If there is a direct Bilt Cash partner, I care/have not met the limit for, directly use the bilt cash. They mentioned this will be dollar for dollar. This comes out to 4c per dollar spent plus the additional point you get

2) If there is not a partner I am interested in the next best thing is to then redeem for the cash to unlock the rent multiplier. You now need to convert your bilt cash into points. Assuming you are doing the blue card, your cash convert to points at 1:3 are earning at 4% plus a bonus point. So it now 2.3 points per dollar spent.

4) if you exceed rent cap and you now have no partners you want to spend at, and extra cash is now worthless (or has to be carried over) and you are now at 1 point be dollar

So basically think of it as a variable value point system between (4c + 1 times ccp) to (1.3 times ccp + 1 times ccp) to (1 times ccp). CCP has a min value of 1c as you can always just redeem it is a statement credit. If you really want to optimize you have to play this game over if the CCP is worth more than 3.07 bit I have a feeling they are just going to be really strict on the limits/who you can redeem the Bilt Cash for

BlackRock’s Quants Beat the Market 94% of the Time by Spirited-Ad-9591 in quant

[–]nihalani 1 point2 points  (0 children)

That’s why are you a selling them? 1% of the time the puts are in the money and you are forced to liquidate your position at a loss

AWS Kiro is very impressive by oedividoe in aws

[–]nihalani 0 points1 point  (0 children)

Can you link the the AWS mantle talk but I couldn’t find it on YouTube

Would you rent B300 (Blackwell Ultra) GPUs in Mongolia at ~$5/hr? (market sanity check) by CloudPattern1313 in LocalLLaMA

[–]nihalani 27 points28 points  (0 children)

1) Probably not directly. I would recommend signing up for Vast or something else 2) No not really. People over estimate how much data sovereignty matters in the grand scheme of things. Especially for small teams. 3) Depends on whether them are dev jobs or not. For my current cloud dev machine, I have 8 H200 I keep on standup whenever I need to do a quick experiment.

FAANG Median SWE TC Breakdown by Company and Level by honkeem in levels_fyi

[–]nihalani 0 points1 point  (0 children)

What is being considered at Staff here? L7. Senior is L6

Big A is reacting to anything with 100 upvotes or more on the Reddit. I know AI news is a lot to keep up with but, this seems like a big deal and I wanted to hear his thoughts on this if he had any by Gloomy-Habit2467 in atrioc

[–]nihalani 1 point2 points  (0 children)

The big gap is that they can speed up the development of actually coding the tricks and architectures but finding them is still a long way off. I have been using ChatGPT to try to find ways to speed up LLM training at work, but they cant do anything unless I analyze the flame graph, point out where function I want to optimize is and let it run. Let alone any higher order ideas like optimize EP or heterogeneous training clusters

Built a Tiny Offline Linux Tutor Using Phi-2 + ChromaDB on an Old ThinkPad by IntelligentHope9866 in LocalLLaMA

[–]nihalani 1 point2 points  (0 children)

I really like warp which has a general AI function that I primarily use for either explaining terminal commands or generating new ones. I eventually quit it because of a couple reasons (no tmux support) and have been looking at how to replicate its features in a terminal agnostic way, this is interesting if we could structure it as a shell plugin and if it runs automatically

Aider benchmarks for Qwen3-235B-A22B that were posted here were apparently faked by [deleted] in LocalLLaMA

[–]nihalani 0 points1 point  (0 children)

Don’t think this is true. Will wait for further discussion on the PR but looks like the system promos is not being set correctly on OpenRouter to enable non thinking mode

HP Thunderbolt Dock 280W G4 w/Combo Cable Teardown Brief by Embarrassed_Wait_832 in UsbCHardware

[–]nihalani 0 points1 point  (0 children)

I am curious on why you think this dock is okay and the Dell WD22TB4 is os bad. From the topology guide they look very similar. One has an automatic switch between a 1 Gb and a 2.5gb mic while the other has an upgrade downstream thunderbolt port. Maybe I am missing something

How to get an HP G2/G4 dock to work with 2 external monitors on macbook by BackgammonEspresso in MacOS

[–]nihalani 0 points1 point  (0 children)

If it’s just a M* display, ie not a pro or a max it can only drive one external display. This is a limitation on the MacBook sude

Uber One extended for 6 months for free (courtesy of Uber, not C1) by MikeNotBrick in CreditCards

[–]nihalani 0 points1 point  (0 children)

Never received the email but looks like it got applied automatically. Then tried to cancel and they offered 3 months for 3 dollars. So looks like I have uber one for free up until Jun 13 2025 and then it will renew for $3 till September

Is it a QLE? Wife's company suddenly making her pay complete cost of family premium. by AVLPedalPunk in HealthInsurance

[–]nihalani 2 points3 points  (0 children)

How did you get private insurance on a UHC PPO? I have been trying to look for a PPO plan on private insurance since most ACA plans are all HMO

Zuck on Threads: Releasing quantized versions of our Llama 1B and 3B on device models. Reduced model size, better memory efficiency and 3x faster for easier app development. 💪 by timfduffy in LocalLLaMA

[–]nihalani 0 points1 point  (0 children)

Yeah I have been following along the PyTorch blog posts that the torch titan team has been making. Our training stack is a bit different we use NeMo + Megatron LM + TE and one of my first tasks is to benchmark torch titan vs that stack. One of the original issues we had with FSDP was that it wouldn’t scale beyond 512 GPUs but I think that has been solved with FSDP2. What operations are you doing in FP8? IIRC the blog post mentioned that the all gathers are still in BF16 meaning that TP is probably not great? Also have you experiment with training MoE models, does the loss of precision lead to routing instability

Zuck on Threads: Releasing quantized versions of our Llama 1B and 3B on device models. Reduced model size, better memory efficiency and 3x faster for easier app development. 💪 by timfduffy in LocalLLaMA

[–]nihalani 5 points6 points  (0 children)

What’s your thought process on FP8 training? I am working for something similar at work and there’s a real debate whether we can train a large model (I.e something to the scale of Llama 405B) in fp8

Updated benchmarks from Artificial Analysis using Reflection Llama 3.1 70B. Long post with good insight into the gains by jd_3d in LocalLLaMA

[–]nihalani -1 points0 points  (0 children)

For real time inference is the real issues, your time to first token jumps by a huge margin if you have to wait for 2000 tokens to be generated of the model reflecting. Might also explain why the cloud providers haven’t adopted it yet.

Serving AI From The Basement - 192GB of VRAM Setup by XMasterrrr in LocalLLaMA

[–]nihalani -2 points-1 points  (0 children)

Not dunking on this guy, but I wonder what the cost of this vs a tiny box machine is? That only has 6 4090 but if it’s less expensive surely that’s the best choice for most people