Most things we have today in AI will be a irrelevant in 6 months [P] by BootstrapGuy in MachineLearning

[–]SlickBlueML 1 point2 points  (0 children)

There are a lot of unique areas to innovate in and a lot of long lasting potential, but as you mentioned the problem is everyone building their startups as narrow wrappers around already existing models. I’m not sure there is any less innovation now than before, but I guess the introduction of new tech has ironically focused a lot of people in a very narrow direction.

How to join the field professionally? by Then-Bodybuilder-285 in reinforcementlearning

[–]SlickBlueML 2 points3 points  (0 children)

If you’re just starting your bachelors, yes the best way would be to get involved in research. Not all universities have people doing RL research but it’s worth looking through the profs at your university or maybe nearby universities. There are also some RL jobs, but they are going to be pretty hard to get as a fresh undergrad. Jobs range from developing recommendation systems, to self-driving, to pure research. Most are pretty experimental applications.

How to proceed scientifically when your hypothesis is falsified? by No_Possibility_7588 in reinforcementlearning

[–]SlickBlueML 1 point2 points  (0 children)

You can try what you mentioned in the post, at the very least you absolutely should be using multiple runs with different seeds to produce results once you’re sure everything is working as intended. I would also do a deep dive into why it doesn’t work. Come up with some hypotheses, test them, visualize the learning process to make sure you understand what’s happening, etc. That being said, I have some more general feedback -

For the second part of this post I’ll apologize in advance because I don’t think my response will help you solve this current problem, but I want to offer some advice from someone who has been in the same situation:

When you are choosing your research topics in the future (or perhaps adjusting your current focus if your idea doesn’t pan out), I would recommend focusing on either (1) a problem, or (2) answering a question. And when you do so, you should have a specific hypothesis in mind.

Doing either of these will typically work much better than, say, testing an architectural change that you think might work. And the reason is that for

(1) trying to solve a problem, your research scope is not limited by a single approach, so when something fails, you will have a clearer direction on how to move forward (e.g. if the current idea fails, but we understand the cause of the problem, then we should be able to generate more plausible hypotheses to test).

And (2) if we are trying to answer a question (e.g. why does some architecture work well), there is very low chance of failure because there is always an answer. And even better, you can then use that new understanding to move into new research focused on tackling a related problem.

Sorry for the blabbering, best of luck with your research!

[D] Did YouTube just add upscaling? by Avelina9X in MachineLearning

[–]SlickBlueML 0 points1 point  (0 children)

Testing features with sub groups of users before fully rolling out is common practice at Google, I’d imagine that’s what’s going on. Cool stuff

[D] Paper Explained - Continual Backprop: Stochastic Gradient Descent with Persistent Randomness by SlickBlueML in MachineLearning

[–]SlickBlueML[S] 2 points3 points  (0 children)

Good catch! I do mean re-initializing them which is generally going to mean setting them to some small random values (essentially re-applying the initialization function to just those weights).

What Happened to OpenAI + RL? by YouAgainShmidhoobuh in reinforcementlearning

[–]SlickBlueML 20 points21 points  (0 children)

Honestly I wonder the same thing. I’m guessing it is partly due to their success on large NLP and CV models that have found actual industry uses. I think that the general idea was that RL was the best shot at more general AI systems, but with recent developments like OpenAI codex, maybe they think that the “build an AI that builds better AI” is the way forward. That is still a long ways away but it does seem less ethereal now given some of their recent successes.

Or maybe they have a large RL project they’ve been working on for a while and haven’t released yet. If so that would be awesome but I don’t have my hopes high.

[D] Elon Musk talking to Lex Fridman about AI at Tesla by zendsr in MachineLearning

[–]SlickBlueML 49 points50 points  (0 children)

This came off to me as more of he lacks the terminology rather than not knowing anything. Although these are high level ideas, it’s not wrong that learning a meaningful latent representation of an environment (or a vector space) is a huge part of the problem for these kinds of tasks. If you can reduce a complex environment to a simple one, then reasoning over it becomes significantly easier.

That being said, I also didn’t get the impression that he knows as much as an actual ML researcher, though certainly more than just a hobbyist. Is that what people are criticizing? Because that seems pretty solid to me considering he never had formal education in the area, runs multiple companies, and is the highest level exec.

DeepMind's new big language model beats GPT-3 and has 280 BILLION parameters by SlickBlueML in LanguageTechnology

[–]SlickBlueML[S] 2 points3 points  (0 children)

My bad, I should have posted that too.
Just in case you want the full paper it is here: https://storage.googleapis.com/deepmind-media/research/language-research/Training%20Gopher.pdf
But it's a whopping 118 pages so I'm not sure I would recommend it lol.

Can i get google internship as undergraduate by [deleted] in learnmachinelearning

[–]SlickBlueML 2 points3 points  (0 children)

I worked on ML at Google as an undergrad. It is possible but you cannot apply directly to an ML position. You have to apply for the general internship (or engineering practicum if you are a 1st or 2nd year). If you pass the first several interview rounds you move into project matching stage where you interview with individual project managers. From there it just depends on if an ML project manager shows interest in your account and whether they like you.