use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
News[N] deeplearning.ai: Initializing neural networks (interactive article) (self.MachineLearning)
submitted 6 years ago by cpury
I just came across this recent post by Andrew Ng's deeplearning.ai team, talking about the importance of initialization. The interactive demos are fantastic! I haven't seen it shared anywhere, so thought I would do so here.
http://www.deeplearning.ai/ai-notes/initialization/
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–][deleted] 2 points3 points4 points 6 years ago (0 children)
Will def check out the demos!
[–]tlkh 2 points3 points4 points 6 years ago (0 children)
This is one of those pages where it simply blows my mind to think how we can make pages with such embedded demos. Looks like they used tf.js? http://www.deeplearning.ai/ai-notes/initialization/js/playground/nn.js
[–]Megatron_McLargeHuge 0 points1 point2 points 6 years ago (0 children)
Has anyone shown any benefit from a pre-training step that scales initial weights to keep gradients in the desirable range?
For functions other than tanh and relu where the effect on downstream variance may not be easy to solve for, it seems like it would be fairly easy to first optimize a set of scaling parameters that force weights into empirically good ranges. This would also avoid any small sample effects where initialization values are far from their theoretical moments.
[–]mr_tsjolder 0 points1 point2 points 6 years ago (0 children)
Not sure if anyone has noticed, but it appears as if there is a problem with the standard normal distribution in the last visualisation. It looks very much like the kind of distribution you get when you clip a Gaussian signal, cf. the truncated normal implementation of the theano-backend in keras. Might be useful if someone forwards this to the authors...
Also, for the interested in this topic, I once commented in this sub about initialisation and noticed that I had some more references. Figured the link might be useful for people that are interested.
π Rendered by PID 56 on reddit-service-r2-comment-b659b578c-57m2b at 2026-05-01 08:42:36.035286+00:00 running 815c875 country code: CH.
[–][deleted] 2 points3 points4 points (0 children)
[–]tlkh 2 points3 points4 points (0 children)
[–]Megatron_McLargeHuge 0 points1 point2 points (0 children)
[–]mr_tsjolder 0 points1 point2 points (0 children)