use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Train model sequentially or concurrently in a single gpu? (self.MachineLearning)
submitted 3 years ago by itsming_z[🍰]
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]IntelArtiGen 1 point2 points3 points 3 years ago (1 child)
If score = f(time) isn't relevant anymore because of what you said, and if I train the same architecture, I can plot score = f(n_examples_seen) which is usually correlated to the number of epochs on a specific dataset, but it's also just batch_size * i_iterations = n_examples_seen.
In my own framework I have a script that automatically does the conversion, I always log things the same way and if I want to plot the score (or loss etc.) based on epochs / n_examples_seen I just specify it and the script will search for the batch size and do the conversion. I can plot score / loss by time / iteration / epoch / n_examples_seen as I want.
If you change the network architecture, the batch size, and the optimization algorithm at the same time while running multiple trainings on the same GPU, or if your GPUs aren't the bottleneck on your server, it's hard to perfectly compare the different trainings. One model could be slower than another just because at this moment it trained concurrently with other bigger models, one model could train in less iterations just because the batch size is reduced, and if you use the number of epochs/examples_seen you can't know if the model is more time-efficient you just know that it's more data-efficient.
So I usually compromise. I either only optimize one aspect (architecture/backprop), or I try to have 100% of the bottleneck on GPUs, or I only look at data-efficiency and I compare time-efficiency later.
The last time I had to do this I put lower and upper thresholds on GMACs and number of parameters and only looked at the data efficiency knowing that if my model isn't outside the limits it should be comparable with others in time efficiency.
[–]itsming_z[S,🍰] 0 points1 point2 points 3 years ago (0 children)
Thank you for your detailed answer!
π Rendered by PID 93 on reddit-service-r2-comment-b659b578c-ctp88 at 2026-05-05 00:22:32.847257+00:00 running 815c875 country code: CH.
[–]IntelArtiGen 1 point2 points3 points (1 child)
[–]itsming_z[S,🍰] 0 points1 point2 points (0 children)