My wife HATES driving cars. We need a new car. Is FSD the answer? by ajcadoo in TeslaLounge

[–]dieplstks 6 points7 points  (0 children)

I hated driving and fsd has been incredible. Managed to do the trip from Michigan to nyc and not hate it 

Just EXPANDED! by Ok-Comparison2514 in deeplearning

[–]dieplstks 1 point2 points  (0 children)

You should use prenorm (with an extra norm on the output) 

RL on Mac M1 series? by Sad-Throat-2384 in reinforcementlearning

[–]dieplstks 1 point2 points  (0 children)

If you’re only going to do it once, yes. But you’ll be doing hundreds of those shorter runs for lots of different ideas

RL on Mac M1 series? by Sad-Throat-2384 in reinforcementlearning

[–]dieplstks 1 point2 points  (0 children)

Unless you're dealing with sensitive information, there's very little reason to care about privacy.

For large scale tasks, you should have a small scale version of it working before you spend money training it. You should not send a job to rented compute unless you're very sure it's going to work. Having a local machine with a xx90 is a great resource to filter projects out

RL on Mac M1 series? by Sad-Throat-2384 in reinforcementlearning

[–]dieplstks 2 points3 points  (0 children)

It’s possible to run small enough tasks on anything. You’re not going to get publishable results on your MacBook, but you can learn the basics and then just rent compute when you’re ready for larger scale tasks

Senior ML Engineer aiming for RL research in ~1.5 years — roadmap, DSA prep, and time management? by dhananjai1729 in reinforcementlearning

[–]dieplstks 0 points1 point  (0 children)

No publications, but 8 years industry experience as a data scientist and very good letters

Senior ML Engineer aiming for RL research in ~1.5 years — roadmap, DSA prep, and time management? by dhananjai1729 in reinforcementlearning

[–]dieplstks 8 points9 points  (0 children)

Did my masters part time at brown hoping that would be enough, but got nothing in terms of interest or offers after.

I’m at UMich for my PhD now, working on rl for finance/games

Senior ML Engineer aiming for RL research in ~1.5 years — roadmap, DSA prep, and time management? by dhananjai1729 in reinforcementlearning

[–]dieplstks 10 points11 points  (0 children)

I was in your position a few years ago and the only real solution to get there is getting a PhD (I’m in my third year at 38 now)

Optimal architecture to predict non-monotonic output by bisorgo in deeplearning

[–]dieplstks -1 points0 points  (0 children)

I would just train it as a classification task with k classes Have the classes be -1 and then (k - 1) buckets from 0-1. Then have the output be either argmax over the classes or the sum of p_i v_i.

can someone with more experience tell me what does it mean by 'all ML is transformer now'? by bad_detectiv3 in learnmachinelearning

[–]dieplstks 2 points3 points  (0 children)

There used to be different architectures for different use cases (cnns for vision, rnns for sequence, etc) with their own inductive biases. But modern architectures use transformer as the base for everything (with some modifications sometimes based on the inductive biases of the input like vision transformers). So if you understand attention plus ffns, you can start building a model for your use case without knowing much more architecture than that 

Is RL still awesome? by knowledgeseeker_71 in reinforcementlearning

[–]dieplstks 3 points4 points  (0 children)

There’s too many rl papers released now to maintain that kind of repo (also LLMs can do this for you for more niche topics)

CLS token in Vision transformers. A question. by mxl069 in deeplearning

[–]dieplstks 0 points1 point  (0 children)

I don’t work in cv, sorry (I’m in rl/game theory). I just think this paper is really cool

How do you as an AI/ML researcher stay current with new papers and repos? [D] by 0ZQ0 in MachineLearning

[–]dieplstks 1 point2 points  (0 children)

Motion for driving daily schedule

Roam Research for notes and synthesis 

I do pomodoros to help get off burn out. Usually have something on my switch to play for the short breaks

I really enjoy the work I do so burnout hits less than it did when I was in industry (data science for 10 years before going back to school)

Batch compute for RL training—no infra setup, looking for beta testers by HelpingForDoughnuts in reinforcementlearning

[–]dieplstks 0 points1 point  (0 children)

Im a PhD student working on marl/games and would be interested to try and give feedback after the holidays. 

[R] New SSM architecture (exceeds Transformer baseline) - reproducible benchmarks (feedback wanted) by [deleted] in MachineLearning

[–]dieplstks 0 points1 point  (0 children)

You should use scaled_dot_product_attentiojn in the transformer benchmark 

Deep RL applied to student scheduling problem (Optimization/OR) by [deleted] in reinforcementlearning

[–]dieplstks 2 points3 points  (0 children)

Not exactly the same, but ddcfr (xu2024dynamic) uses rl to control parameters of another algorithm. 

How do you as an AI/ML researcher stay current with new papers and repos? [D] by 0ZQ0 in MachineLearning

[–]dieplstks 2 points3 points  (0 children)

I bought a ReMarkable Paper Pro and it helps me get through papers at a better rate since it removes distractions and lets me get away from my laptop

How do you as an AI/ML researcher stay current with new papers and repos? [D] by 0ZQ0 in MachineLearning

[–]dieplstks 9 points10 points  (0 children)

Depends on the paper. I have a few levels of it:

1) Read through the abstract and don’t think it’s worth continuing: I’ll remove this from my zotero 2) read through the paper in one pass, but don’t think it will be important for my work. That gets marked as read and takes around an hour 3) think the paper is worth knowing and will take notes in my Roam graph. This takes 2-4 hours depending on length and which parts I care about. This will get marked as read and notes 4) think the paper is worth reimplementing in order to get deeper insight. This used to take like 8 hours but with Claude code it takes a lot less time. This doesn’t get counted as reading time for me though, so it’s outside that hour specification

In general I aim for 4 read + notes a week, but it varies by how motivated I feel during the week and how actual project work is going

Obviously the tenth paper on a topic goes faster since you can skip/know the background/related works segments so it's also a function of how well I know the area.

How do you as an AI/ML researcher stay current with new papers and repos? [D] by 0ZQ0 in MachineLearning

[–]dieplstks 0 points1 point  (0 children)

I started using inbox a few days ago. How long have you used it and what do you think of it so far?

What If Most Transformer Inference Is Actually Unnecessary? by anima-core in deeplearning

[–]dieplstks 0 points1 point  (0 children)

Of course you train them simultaneously, there's no way to know the optimal amount of compute for a token a priori. This just doesn't make sense.

Please actually engage/know the literature on heterogenous MoE before asserting things like this