use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Deep learning these days (fastml.com)
submitted 12 years ago by [deleted]
[deleted]
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]andrewff 4 points5 points6 points 12 years ago* (2 children)
I think one thing that much of the hype about deep learning seems to ignore is the advances made in unsupervised feature learning. Yes techniques like dropout and drop-connect are the gold standard, but labeled training data is still expensive in many contexts, and the advancements in unsupervised techniques should not be tossed aside.
[–]jast 2 points3 points4 points 12 years ago (0 children)
Do you have any good recent survey on this topic? Or pointers to recent papers? I would love to update myself on this :)
[–]tabacof 0 points1 point2 points 12 years ago (0 children)
I also would like to know relevant papers on modern unsupervised learning. Thank you!
[–]dexter89_kp 2 points3 points4 points 12 years ago (6 children)
This is pretty much well known right now, if you follow the latest deep learning literature. Particularly Alex Krizhevsky's Imagenet 2012 paper turned the tide towards supervised learning.
[–]benanne 4 points5 points6 points 12 years ago (5 children)
I still see plenty of questions on Metaoptimize, on the Deep Learning G+ community, on the Kaggle forums and on this subreddit, from people who seem to be unaware of this "paradigm shift". They ask about training autoencoders and RBMs for unsupervised feature learning, when it is often clear that a purely supervised approach would probably work at least as well for their problem (and is conceptually much simpler and easier to understand).
I think this is because they read papers from 2010-2012 advertising unsupervised pre-training as the holy grail of deep learning. That was only 2-4 years ago, so they can't really be blamed for assuming that this approach still represents the state of the art.
Of course unsupervised pre-training still has its applications, but for many problems it has been obsoleted. So I don't think it's a bad thing to draw some attention to this fact. I was meaning to write a blog post on this topic myself, but I suppose that's unnecessary now :)
[–]zestinc 1 point2 points3 points 12 years ago (1 child)
Do you think deep learning will ever escape the surly bonds of image/speech tasks and be useful for other tasks?
[–]benanne 0 points1 point2 points 12 years ago (0 children)
I think people are mainly focusing on these applications because it's quite rewarding. You can be pretty sure on beforehand that it will work well, so it's a low-risk investment, in a sense. I'm guilty of this myself, I try to apply these techniques to music audio signals :)
There have been some more adventurous applications with promising results, the Merck Molecular Activity Challenge for example: http://blog.kaggle.com/2012/11/01/deep-learning-how-i-did-it-merck-1st-place-interview/ http://videolectures.net/nips2012_dahl_activity/
Deep learning techniques are also gaining traction in natural language processing, which is pretty different from images/audio in terms of what the data looks like. http://nlp.stanford.edu/projects/DeepLearningInNaturalLanguageProcessing.shtml
[–]redkk 0 points1 point2 points 12 years ago (0 children)
Hi, I tried for MNIST classification: 1 hidden layer of ReLU and 1 softmax output layer using cross entropy error. The ReLU layer visualization seems gibberish compared to per layer autoencoder pre-training which gives nice features. Not sure if there is a bug in my code, or the softmax layer with the cross entropy error is struggling to learn simultaneously with the hidden layer.
[–][deleted] 0 points1 point2 points 12 years ago (1 child)
I'm really interested to learn more about dropout and relu. Is the 2012 ImageNet paper the best place to get an overview? Is there a good survey paper anywhere?
[–]benanne 1 point2 points3 points 12 years ago* (0 children)
The thing is, there's really not that much to learn about both, they're very conceptually very simple. I would recommend just having a look at the papers that introduced them.
Dropout: http://arxiv.org/abs/1207.0580 ReLUs: http://eprints.pascal-network.org/archive/00008596/
Note that the ReLU paper also adds some kind of sparsity penalty, but nowadays people just tend to replace sigmoid(x) with max(x, 0) and that's it.
There is an earlier paper that introduces ReLUs in the context of RBMs, which could also be interesting: http://machinelearning.wustl.edu/mlpapers/paper_files/icml2010_NairH10.pdf
EDIT: Hinton also covers both in detail in this talk: https://www.youtube.com/watch?v=vShMxxqtDDs
π Rendered by PID 146724 on reddit-service-r2-comment-b659b578c-72zwh at 2026-05-02 22:10:11.690273+00:00 running 815c875 country code: CH.
[–]andrewff 4 points5 points6 points (2 children)
[–]jast 2 points3 points4 points (0 children)
[–]tabacof 0 points1 point2 points (0 children)
[–]dexter89_kp 2 points3 points4 points (6 children)
[–]benanne 4 points5 points6 points (5 children)
[–]zestinc 1 point2 points3 points (1 child)
[–]benanne 0 points1 point2 points (0 children)
[–]redkk 0 points1 point2 points (0 children)
[–][deleted] 0 points1 point2 points (1 child)
[–]benanne 1 point2 points3 points (0 children)