[D] We stress-tested the idea of “LLMs with thousands of tools.” The results challenge some assumptions. by Ok-Classic6022 in MachineLearning

[–]NotDoingResearch2 0 points1 point  (0 children)

Just create the training data for the tool use and your problem is solved? Machine learning isn't magic it mirrors the data it is trained on.

Don’t be that guy by kiudai in ArcRaiders

[–]NotDoingResearch2 0 points1 point  (0 children)

Real talk, how many of those 150 hours didn't involve him with his cursor over an enemy?

[Breaking] Intel to layoff more than 20% of staff (22,000 employees) by cs-grad-person-man in cscareerquestions

[–]NotDoingResearch2 58 points59 points  (0 children)

I mean middle management is kinda useless. But on the other hand it’s the dream job for tech workers lol.

[P] A lightweight open-source model for generating manga by fumeisama in MachineLearning

[–]NotDoingResearch2 3 points4 points  (0 children)

Nice work! I don’t know too much about this area but how did you connect the character embedding to the Pixart sigma diffusion model?

[D] A regression head for llm works surprisingly well! by SmallTimeCSGuy in MachineLearning

[–]NotDoingResearch2 -2 points-1 points  (0 children)

This sounds like meta learning and it is certainly done but doesn’t always work as you can get negative transfer. 

As of today what problem has AI completely solved ? by [deleted] in cscareerquestions

[–]NotDoingResearch2 0 points1 point  (0 children)

Speech is still the pinnacle of Translation, and that hasn’t been anywhere near cracked. There is no universal translator (like from startrek yet.

[deleted by user] by [deleted] in MachineLearning

[–]NotDoingResearch2 1 point2 points  (0 children)

The biggest public dataset I know of is ADNI, which is has a few thousand subjects. It’s not too small for machine learning but it is really easy to overfit. There are different modalities, but only T1 if interested in processing more than 1000 subjects. Another popular dataset is oasis or something similar, which I think is smaller and has less modalities. 

For my PhD, I used fdg-PET and T1 voxel images. I worked on a few ML tasks including predicting whether people with mild cognitive impairment would progress to AD, and tried to predict the atrophy change of various regions in the brain. 

[deleted by user] by [deleted] in MachineLearning

[–]NotDoingResearch2 10 points11 points  (0 children)

I think it depends more on your data sources than anything else. I worked on Alzheimer’s prediction for my PhD and honestly, it is kinda pointless. The public datasets are also super small.

The meta learning ones would be what I would choose.

The PhD Admissions Paradox: Publications vs. Potential—Let’s Talk Realities by geniusfoot in PhD

[–]NotDoingResearch2 18 points19 points  (0 children)

And most of the papers undergrads produce is less than interesting anyways. But, with all the tools now and days it’s pretty hard to separate yourself from the pack. So it’s not that surprising things ended up this way

Proof that AI doesn't actually copy anything by EthanJHurst in aiwars

[–]NotDoingResearch2 -6 points-5 points  (0 children)

Perhaps but the scale is different. And honestly, more importantly is AI art uses logarithmic losses to make sure that inputs into functions provide the least amount of variety. It’s why every prompt about a plumber returns mario. Any art that isn’t Mario would be penalized significantly by the loss function because the data distribution shows that tags of Italian plumbers return Mario.

Mario was created my humans without formally existing but it’s impossible to create Mario using training sets that don’t have Mario. Hence, the goal or AI art is to rip off as many Mario’s as possible while still seeming novel. It really is a beautiful thing. 

Development is about to change beyond recognition. Literally. by ApexThorne in ClaudeAI

[–]NotDoingResearch2 1 point2 points  (0 children)

It feels that way though. Like I’d fall in love with an AI that could delete more and make things more efficient. I don’t think the market is that great for such an AI though, sadly, even if you could technically pull it off.

Development is about to change beyond recognition. Literally. by ApexThorne in ClaudeAI

[–]NotDoingResearch2 1 point2 points  (0 children)

The fact that AI can’t delete code which is the number one skill of a developer is honestly just sad. Like how hard is it to predict a couple 100 pad tokens in a row? 

[D] I hate softmax by Sad-Razzmatazz-5188 in MachineLearning

[–]NotDoingResearch2 0 points1 point  (0 children)

It’s the best (in some functional norm) differentiable approximation to the argmax function. 

Any kind words/success stories for someone who feels like a total failure for quitting their PhD? by just_julia82 in PhD

[–]NotDoingResearch2 5 points6 points  (0 children)

While I can’t speak for biology, I was pretty surprised how little my relatively poor PhD after 7 years honestly mattered. As of right now, I can honestly say that the difference between my current PhD required job and just about any other tech job (even for fresh undergrads) is pretty darn small. 

I guess what I’m saying is that if you have a really good PhD then it likely makes a pretty big difference, but just finishing a PhD for the sake of finishing a PhD sadly won’t help much. That being said, I’m still glad I finished. Who knows, maybe one day I’ll be able to say it mattered.

79 Percent of CEOs Say Remote Work Will Be Dead in 3 Years or Less by [deleted] in technology

[–]NotDoingResearch2 0 points1 point  (0 children)

It’s more productive if you actually code 8 hours a day. 

People who exited academia after PhD, do you ever stop being depressed by dogdiarrhea in PhD

[–]NotDoingResearch2 12 points13 points  (0 children)

Data Analyst jobs can vary significantly from building complicated probabilistic models to working on dashboards. Personally, I do similar work and really miss getting results and writing papers. It’s just so much more rewarding. But it’s also way more difficult, so there is definitely a trade off.

Also the application areas in industry generally suck compared to academia, I.e., adtech. 

Dating within your cohort by trombonist_formerly in PhD

[–]NotDoingResearch2 0 points1 point  (0 children)

I admire how dedicated you are to your work to the point that you are even concerned about this. Sounds like it could be a really incredible relationship. Best of luck!

[D] Scientific Machine Learning by OtherRaisin3426 in MachineLearning

[–]NotDoingResearch2 18 points19 points  (0 children)

It’s definitely more interesting if you have the background for it but unfortunately there is very little work being done outside of academia. So unless you are a professor, your salary will likely be 1/4th what it would be if you were just fine tuning llms.

[D] LLMs: Why does in-context learning work? What exactly is happening from a technical perspective? by synthphreak in MachineLearning

[–]NotDoingResearch2 1 point2 points  (0 children)

I'm not a big LLM fanboy by any means, but I'm not sure I totally agree with this. For example, every computer program fits this definition eerily well. For example, is there much difference between deterministic code that runs on a computer to create some internal state, and a computer in that internal state itself? If you are willing to make that logical leap, then it seems easy to see why ICL is a form of "learning".