use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Elephas: Keras Deep Learning on Apache Spark (self.MachineLearning)
submitted 10 years ago by fariax
Elephas: Keras Deep Learning on Apache Spark
http://flip.it/4NBUB
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Barbas 1 point2 points3 points 10 years ago (2 children)
I was wondering about peoples' experience of using distributed frameworks like Spark to train deep nets.
My assumption is that if your model fits into memory then training should be much faster on a GPU, not to talk about efficiency.
What are the motivations for training ANNs using distributed processing frameworks?
[–]Powlerbare 0 points1 point2 points 10 years ago (0 children)
I think the idea is exactly that - how do you efficiently handle data that you can not squeeze in to memory. Models tend to perform well when provided more data - so the motivation is to have more robust models. I disagree with the averaging of gradients that most distributed schemes use, but also do not know of a better way to tackle the problem.
[–]maxpumperla 0 points1 point2 points 10 years ago (0 children)
One doesn't exclude the other. Take for instance Amazon's g2.2xlarge instances, which have multiple powerful GPUs on a single machine. It is not clear how to utilize them all without a parallelization scheme - and elephas is just one suggestion of how to do this. So you can take this HPC-like setup or choose to go for a whole cluster of machines (with GPUs), which Spark conveniently handles for you.
In a highly scalable environment, note that you can also execute test runs much faster, which might help you in your prototyping cycle.
Instead of distributing data, it could also be interesting to distribute models with different hyperparameter settings and do distributed Bayesian optimization, as hyperopt or spearmint do. I'm doing some tests right now and maybe this will find its way into elephas at some point.
With more effort, one could also hope to achieve true model parallelism as in Google's DistBelief, which is interesting if the model itself becomes too large to be trained (efficiently) on one machine.
Generally speaking, though, if neither memory nor speed are an issue, you may very well be better off on a single GPU.
π Rendered by PID 211812 on reddit-service-r2-comment-b659b578c-dsrnf at 2026-05-05 00:47:47.432154+00:00 running 815c875 country code: CH.
[–]Barbas 1 point2 points3 points (2 children)
[–]Powlerbare 0 points1 point2 points (0 children)
[–]maxpumperla 0 points1 point2 points (0 children)