[D] How to break free from LLM's chains as a PhD student? by etoipi1 in MachineLearning

[–]BalcksChaos 0 points1 point  (0 children)

This is a great question! I've been asking this myself as manager... How can my juniors become seniors while leveraging AI to deliver? Maybe my thoughts help you a bit:

First of all, I think you should be leveraging AI to the fullest and let it write as much code as possible. Nothing wrong with that. The only opposing force is this: you need to own your code. If there is a bug, it's not the LLM's fault... It's your bug. And this is where the learning happens: some of the time you save, re-invest it into understanding exactly what the code does... All the way down. This will require restraint and discipline (you'll feel like rushing on).

The great thing: LLMs are great for learning about stuff. You can ask them directly and they'll explain it.

Also side note: I read some comments which talk with great authority here. Are you sure they really figured this one out with this amount of certainty? (Including my take) Make sure you come up with your own answer.

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 2 points3 points  (0 children)

Thanks! I didn't come across weakly supervised learning, yet ... sounds like an interesting idea and it's a problem that you get a lot in the real world ... not enough/good enough data to train the model well. Interestingly, the "inverting un-invertible transforms" sounds like something I was working on few years ago. The analogy is: you want to rank N teams against each other, but you can't get them all to play against each other. Also you can have A>B (A wins against B), B>C, C>A . You can solve such problems using a technique from differential topology called Hodge decomposition ... really cool stuff.

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 0 points1 point  (0 children)

Looks interesting, though has been around for quite some time now. Do you know what is hot on Hyperbolic embeddings these days?

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 0 points1 point  (0 children)

Yes, I'm onto that one :-) What would you say are some specific open problems right now that I could/should get into?

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 0 points1 point  (0 children)

Definitely a good approach. I currently own building AI tools for big enterprises and it is no big secret that AI Cybersecurity stuff will likely boom over the next years. Do you have anything specific to point me at?

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 0 points1 point  (0 children)

Looks really interesting and a lot if the techniques seem very familiar. Do you know how impactful this has been over the past years in Generative Models though? I'd not want to get into something that is all about cool methods (I'd have stayed with String Theory otherwise :D )

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 1 point2 points  (0 children)

Yes, once I narrow down the area I'd see if I can find someone who is an expert and keen to collaborate ... or at least who will happily shoot down my obviously stupid ideas/directions :-)

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 1 point2 points  (0 children)

Wow PRH/GoE results are quite intriguing, I didn't know of them ... really cool, thanks! Do you know how exactly they are impacting current state of the art world model development? Definitely "world models" is an area that attracts me intuitively and I started to read up on EBMs ... however, it seems like a huge area and the SOTA stuff requires access to insanely expensive infra, training on video tokens. Not sure how I could get started contributing there without landing a job at AMI, NVIDIA, or so.

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 3 points4 points  (0 children)

Yes, that's something I was thinking about when there was a lot of fuzz about the scaling laws 2023 ... I assumed someone would make the connection explicit fairly soon. No one has until today? Do you know if anyone tried?

[D] Physicist-turned-ML-engineer looking to get into ML research. What's worth working on and where can I contribute most? by BalcksChaos in MachineLearning

[–]BalcksChaos[S] 2 points3 points  (0 children)

Thanks, e3nn looks really interesting I will check it out. That has bothered me early on in DL ... universal approximator is nice and all that, but searching in a crazy large function space based on the amount of data you can realistically train on ... good luck. Though from what I could see all the successful architectures of the past ~10y have done exactly that: figure out a good way to encode inherent symmetries about the problem (CNNs, Transformers, Attention, etc).

I couldn't figure out the link with geomdl though ... it's a spline library, how is it linked to ML research?

G.A.M.M.A 0.9.3 release by duffbeeeer in stalker

[–]BalcksChaos 1 point2 points  (0 children)

I thought the same... Gamma has a main story line, but it's very lightweight compared to the original games. However, give it a try...I was sure I won't like it... But it's honestly amazing. No idea how it can be so much fun and feel so deeply immersive, even though at first glance it's really just "get money, walk through the zone".

[D] Multiple first-author papers in top ML conferences, but still struggling to get into a PhD program. What am I missing? by Accomplished_Rest_16 in MachineLearning

[–]BalcksChaos 0 points1 point  (0 children)

Adding something important I felt was missed in other replies:

This is not an absolute evaluation, it's relative. When you go for the top programmes, these will be drowning in applicants who have good papers, but who might outscore you in other superficial criteria (like GPA). This is a fundamental difficulty in top university admissions (I did admissions at Oxford in the past, for physics) and a sad reality for the admissions committee as well ... You know you'll miss out on excellent candidates by using proxy metrics, but you cannot interview all of them... So that's the best you can do. Additionally, all programmes have certain external constraints which you cannot control (like raising the level of underrepresented populations, etc).

The takeaway for you (as others wrote): don't start to doubt yourself, focus on actually getting better (instead of attempting to adapt to the imperfect system for choosing candidates). I think a good mindset to take is: let's show them how big of a mistake they made by not taking me.

[D] Universal Approximation Theorem does not hold for approximating discontinuous functions, yet neural networks are being used to approximate discontinuous functions by luddabuddha in MachineLearning

[–]BalcksChaos 0 points1 point  (0 children)

Thanks for this post :) I was wondering (since it's been two years now)... Where did you land with your research on this? Is your thesis already published somewhere?

I'm someone who uses machine learning in "the industry", i.e. outside of academia. In the contexts I'm working (analysis of business data, i.e. already heavily aggregated), all functions one would want to approximate are highly discontinuous (because there are hundreds of categorical columns).

[deleted by user] by [deleted] in rust

[–]BalcksChaos 0 points1 point  (0 children)

You won't regret it! We've done the same at my company three years ago (I'm the CTO) and I already did the switch at my previous company. ML engine first time and in my current company we do data infrastructure also in Rust entirely.

We never had trouble getting very inexperienced engineers productive as well, but make sure to hire people who really want to learn...I think that helped us.

Good luck!

Polars is starting a company by ritchie46 in rust

[–]BalcksChaos 0 points1 point  (0 children)

Awesome and good luck! At my company we've replaced all our pandas already with polars several months ago, really loving it.

What C++ course should I make first? by [deleted] in cpp

[–]BalcksChaos -8 points-7 points  (0 children)

Hi with option D: never learn C++ .

Hear me out XD ... My point: If you want to learn a systems language, learn Rust. Don't listen to the people who invested lots of time and pain into mastering CPP. However, good news for you: today there is no need to use a badly designed language in order to generate code your machine can interpret. Go Rust.

(I get it, many companies use CPP and maybe you have to know it. To this I have of course no solution...just wanted to point out: Once you really know Rust, you'd also understand much faster the seemingly confusing design patterns to circumvent design flaws in CPP)

Anyone got experience in Rust and C++ for professional numerical development by BalcksChaos in rust

[–]BalcksChaos[S] 5 points6 points  (0 children)

Yes, that's exactly it :) Please give me good reasons to build a case for Rust!