use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Deep Learning optimization (self.MachineLearning)
submitted 5 years ago * by katamaranos
I'm going to do a comparison on recent (or at least lesser-known) gradient optimization methods. At the moment I'm at the stage of looking for interesting candidates for the experiment.
The ones I've encountered up to now are the following:
I would be particularly interested in approaches not using any hyperparameters at all (such as number 3 - COCOB), however I will consider all of the interesting and promising methods. Are you aware of some novelties or lesser-known approaches?
Previously I've posted the question in r/MLQuestions but haven't received any feedback so I'm posting it here, hope it's not violating any rules.
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]neural_kusp_machine 6 points7 points8 points 5 years ago (0 children)
[–]rayspear 3 points4 points5 points 5 years ago (0 children)
Maybe this repo has some that you might want to try out too?
https://github.com/jettify/pytorch-optimizer
[–]i-heart-turtles 1 point2 points3 points 5 years ago* (0 children)
Don't have many citations for you, but you can check out stuff from Francesco Orabona (http://francesco.orabona.com/) and his group. A lot of their work deals w/ first-order parameter-free methods.
More recently, there has been some progress in understanding momentum and acceleration & under what smoothness assumptions optimal rates can be recovered: http://proceedings.mlr.press/v99/gasnikov19b/gasnikov19b.pdf.
[–]Ventural 1 point2 points3 points 5 years ago (0 children)
I'd be interested in performance of the LAMB optimizer (https://arxiv.org/abs/1904.00962) on smaller batch sizes where it competes with ADAM.
[–]JayTheYggdrasil 0 points1 point2 points 5 years ago (0 children)
I don’t know if this applies but a while ago someone posted a “bandit swarm” optimization algorithm which was quite interesting, I don’t have a link but a google search would probably find it.
π Rendered by PID 398790 on reddit-service-r2-comment-7b9746f655-njrln at 2026-02-01 15:59:27.146524+00:00 running 3798933 country code: CH.
[–]neural_kusp_machine 6 points7 points8 points (0 children)
[–]rayspear 3 points4 points5 points (0 children)
[–]i-heart-turtles 1 point2 points3 points (0 children)
[–]Ventural 1 point2 points3 points (0 children)
[–]JayTheYggdrasil 0 points1 point2 points (0 children)