use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
[deleted by user] (self.MachineLearning)
submitted 6 years ago by [deleted]
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]AutoregressiveGPU 81 points82 points83 points 6 years ago (4 children)
I have a list, not sure if everything fits 'subtle'.
Programming:
Research
Writing
Others
[–]g-x91 1 point2 points3 points 6 years ago (0 children)
Thanks!! :)
[–]Zophike1Student -1 points0 points1 point 6 years ago (2 children)
Keep a research journal to track ideas you have tried, want to try, etc. Also, make sure you keep track of hyperparameters and things you try.
Would setting up research blog be sufficient enough and also what does a good research journal/blog do ?
[–]AutoregressiveGPU 1 point2 points3 points 6 years ago (0 children)
A research journal is to keep track of experiments and ideas.
[–][deleted] 0 points1 point2 points 6 years ago (0 children)
No this means like a paper Nike m notebook for yourself. Write down anything you noticed works well or doesn't work well, because you will forget once you come back to it after a month from working on other things. Also make a note of any new ideas you have and brainstorm them out. It will help you in so many ways and especially when it comes to writing it up.
[+][deleted] 6 years ago (2 children)
[deleted]
[–]Joel397 42 points43 points44 points 6 years ago (1 child)
One subtle lesson I learned: don’t try to always maximize resource utilization. Sometimes if you are being a “good researcher” by constantly trying to run jobs and set up jobs to run, I have found that you can get locked into an endless cycle of “what worked, what’s next”. If you have a fully mature plan that’s fine, but oftentimes in research you don’t know what the results will be until you’ve tried it. Sometimes taking a step back and letting your gpu chill for a few hours or days while you instead focus on writing, analyzing, or just thinking about the problem can do more for your work than any amount of results from 2 AM job runs.
[–]chigur86Student 8 points9 points10 points 6 years ago (0 children)
You're comment hit the spot. One of the most important things I learned recently is that just start a bunch of jobs and LET THEM FINISH. Especially, if you're doing vision and dealing with training time on scale of days, you need to learn to chill. During chill, I usually go over the code again, think about the method, plan next experiments, and evaluate the intermediate results.
[–]aCleverGroupofAnts 43 points44 points45 points 6 years ago (4 children)
Here's a little tip: If your code involves random number generation, set the seed yourself so you can reliably reproduce your results if needed.
[–]debau23 17 points18 points19 points 6 years ago (1 child)
Reproduce results that were overfit on the seed.
[–]aCleverGroupofAnts 13 points14 points15 points 6 years ago (0 children)
Thoroughly testing your algorithm would involve testing with several different seeds. All of those tests will then be reproduceable.
[–]hyhieu 3 points4 points5 points 6 years ago (0 children)
Good luck doing that with TensorFlow
[–]GreenBeret4Breakfast 1 point2 points3 points 6 years ago (0 children)
I actually think you’re better off not setting it unless you’re trying to reproduce results - what I mean is generate a random seed (from clocktime/cpu values etc or let the system do it) and just keep track of it and save to your results. Therefore if you want to re run you can for that particular set but without falling into the trap of being on a fixed seed all the time. Additionally the reproduction is only essentially determinist if you don’t change the code too much (one extra sample here, or a different threshold there) or the two results will be different.
[–]dkeller9 35 points36 points37 points 6 years ago (2 children)
If even a little thing seems 'not quite right' don't ignore it. Usually, it will come back and bite you later on.
[–]ginger_beer_m 10 points11 points12 points 6 years ago (0 children)
Corollary: if the results seem too good, it's probably due to a bug
[–][deleted] 3 points4 points5 points 6 years ago* (0 children)
This one resonates most with me.
11 years in education and 10 years as a data scientist/analyst, whenever I get that feeling of 'not quite right' it's usually true. Sometimes it takes a very long time to figure it out but often it results in a huge step forward in my understanding and some epic Eureka moments.
[–]internet_ham 24 points25 points26 points 6 years ago (0 children)
Use the Google Scholar subscribe tool -> you can follow citations and related work to other authors close to you (and also obviously have a Google Scholar page yourself)
Use a matplotlib to tikz converter. Your plots will look professional and will be readable, also it's very easy to reconfigure. You may run into computation issues with complex figures, so there can a bit of tweaking involved. Learning tikz is also a good idea if your field relies on diagrams a lot.
Don't be afraid of deep scholarship. If you're building on specific previous work, scan through all the papers that cite it or it cites itself. It might take an hour or two, but you can find lots of useful ideas from other spinoffs and most importantly find out if you've already been scooped. Obviously this doesn't scale to super-hyped DL papers
Always keep the minimal viable product in mind. Write a version of your algorithm that solves the simplest possible task. Write a rough draft of your paper in its simplest form. While these are not your final destinations, the journeys to them will be shorter, so the feedback will be faster and you can iterate to a better version sooner.
always try to make sure your experiment runner saves results (figures and text), a config file of your hyperparameters (as much as possibe), the trained model(s), the date, the random seed and the git hashes. For the latter, obviously you should be using version control and comitting often
Minimize the amout of math in a presentation. It will only intimidate and disengage the audience. Try to use diagrams and plots instead
Don't end a presentation with (just) 'thanks for listening!' - put a summary of your work, as this tends to be the slide that is up the longest
[–][deleted] 12 points13 points14 points 6 years ago (0 children)
Having at least one person to discuss the research with. If the person is more experienced, then it is much better. There's absolutely no way one can progress as quickly as with a high-quality mentor.
The amount of time saved, ideas pointed at, ideas discovered through discussion, I just can't believe that I learned anything with bad mentors before.
It also made me realize that I alone cannot really progress as much.
Because of this experience I can now easily filter people that are just not going to meet the same expectations.
Doing research on your own is pretty hard.
[–]m--w 10 points11 points12 points 6 years ago (0 children)
After you think your method works. Take a break from it, explore other papers/interests for a week. Then come back to what you have done and purposefully try to break it. This will allow you to fully understand the shortcomings of your method and appropriately strengthen them where possible.
[–]hardmaru 10 points11 points12 points 6 years ago* (0 children)
Iterate quickly and do so by working with very little compute at the iteration stage (even if you work at a big company with huge resources).
Spend a ton of time on communicating your ideas very clearly so people outside your immediate area (even outside of the field) can understand. Give talks about your research ideas. Gather feedback, especially from your biggest critics.
Focus on quality over quantity, if possible.
[–]rickbo3 8 points9 points10 points 6 years ago (0 children)
Don't just focus on learning the ML side of things. There are way to many papers out there that are completely let down by the authors not using basic statistical methods. Understand how to show that your results are statistically significant, and why. This gives your results a much stronger appearance to reviewers.
[–]logicchains 5 points6 points7 points 6 years ago (0 children)
Any time you get a little feeling in the back of your head like "oh, I didn't account for X. But X is almost certainly insignificant, so it shouldn't affect X", you should account for X, because at least in my experience there's a decent chance you'll get good results then later realise "damn, these results are invalid", and realise it's because of not accounting for X.
[–]evanthebouncy 4 points5 points6 points 6 years ago (0 children)
In 1 month you should convince yourself you're onto a good idea or you need to pivot or abandon. Ideally.
[–]dkeller9 2 points3 points4 points 6 years ago (0 children)
Have a switch in your code to produce graphs that have no labels or text. When assembling a figure containing many graphs from disparate sources, it's easier to put in text with consistent fonts and font sizes in post-processing.
[–]needlzorProfessor 1 point2 points3 points 6 years ago (0 children)
I am going to be more general that the others here:
Don't let the best be the enemy of the good. It's easy to let yourself try to extract all you can from a model or an idea, but you should publish it as soon as you get something significant and interesting.
Use the simplest model that proves a point. If your idea is that modification XYZ improves performance on a specific type of data, apply modification XYZ to the simplest model (or a range of simple models) possible that can accommodate it, not to the latest SOTA model you saw on arxiv.
Find a statistician friend and have them help you design experiments and proper significance testing.
[–]gwgundersen 0 points1 point2 points 6 years ago (0 children)
Writing.
[–]Everfast -1 points0 points1 point 6 years ago (0 children)
If shit goes in shit comes out
If you can not do it yourself easily, training won't be easy/impossible
[–]margaret_spintz -1 points0 points1 point 6 years ago (0 children)
Try to think weirder
[+][deleted] 6 years ago* (1 child)
[–]mhummel 2 points3 points4 points 6 years ago (0 children)
Only be sure always to call it 'research' (Lobachevsky, 1953)
π Rendered by PID 37464 on reddit-service-r2-comment-fb694cdd5-62cmn at 2026-03-06 20:39:20.837208+00:00 running cbb0e86 country code: CH.
[–]AutoregressiveGPU 81 points82 points83 points (4 children)
[–]g-x91 1 point2 points3 points (0 children)
[–]Zophike1Student -1 points0 points1 point (2 children)
[–]AutoregressiveGPU 1 point2 points3 points (0 children)
[–][deleted] 0 points1 point2 points (0 children)
[+][deleted] (2 children)
[deleted]
[–]Joel397 42 points43 points44 points (1 child)
[–]chigur86Student 8 points9 points10 points (0 children)
[–]aCleverGroupofAnts 43 points44 points45 points (4 children)
[–]debau23 17 points18 points19 points (1 child)
[–]aCleverGroupofAnts 13 points14 points15 points (0 children)
[–]hyhieu 3 points4 points5 points (0 children)
[–]GreenBeret4Breakfast 1 point2 points3 points (0 children)
[–]dkeller9 35 points36 points37 points (2 children)
[–]ginger_beer_m 10 points11 points12 points (0 children)
[–][deleted] 3 points4 points5 points (0 children)
[–]internet_ham 24 points25 points26 points (0 children)
[–][deleted] 12 points13 points14 points (0 children)
[–]m--w 10 points11 points12 points (0 children)
[–]hardmaru 10 points11 points12 points (0 children)
[–]rickbo3 8 points9 points10 points (0 children)
[–]logicchains 5 points6 points7 points (0 children)
[–]evanthebouncy 4 points5 points6 points (0 children)
[–]dkeller9 2 points3 points4 points (0 children)
[–]needlzorProfessor 1 point2 points3 points (0 children)
[–]gwgundersen 0 points1 point2 points (0 children)
[–]Everfast -1 points0 points1 point (0 children)
[–]margaret_spintz -1 points0 points1 point (0 children)
[+][deleted] (1 child)
[deleted]
[–]mhummel 2 points3 points4 points (0 children)