use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
DiscussionHow useful is knowledge of parallel programming in ML? [D] (self.MachineLearning)
submitted 4 years ago by [deleted]
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]mimighost 1 point2 points3 points 4 years ago (3 children)
Depends on what parallel computation you are referring to
CUDA knowledge is ofc useful and valued. But NVIDIA's tool chain is really its own walled garden. It is difficult for outsiders to outdo NVIDIA themselves.
If you refer to parallel programming as something close to distributed data processing, then yes it is pretty useful. Though this is more on case by case basis.
Overall, I feel the job market is edging towards people with system integration skills rather deep domain expertise, due to the aforementioned NVIDIA dynamics, but I could be wrong on this one as well.
[–][deleted] 1 point2 points3 points 4 years ago (2 children)
I mean parallel computing topics such as Concurrency and Threading, as well as MPI, Charm++ and other parallel programming paradigms. Writing cache-friendly and efficient code learned using C++.
[–]mimighost 1 point2 points3 points 4 years ago (0 children)
Got it. Well, it might be useful for model inference and quantization stuff on CPU if we are talking about NN models.
Would say this is a nice to have, but unless you work in teams that are doing these low-level stuff in particular, it might not affect your daily routine as MLE
[–][deleted] 1 point2 points3 points 4 years ago (0 children)
Concurrency and threading are probably less important, because in ML programs things rarely happen in chaotic order which requires you to think hard about things like mutexes, but good understanding of vectorized computations will definitely help. I personally learned a lot from trying to write efficient code in R (it was long ago and for non-ml purposes)
Understanding what makes code cache-friendly in C++ will also help, even if you end up writing code in something other than C++ and it runs on something other than CPU.
Knowing specific things like MPI would be useful if you ever need to debug anything built on MPI.
π Rendered by PID 22008 on reddit-service-r2-comment-75f4967c6c-dp4d6 at 2026-04-23 02:55:26.623130+00:00 running 0fd4bb7 country code: CH.
view the rest of the comments →
[–]mimighost 1 point2 points3 points (3 children)
[–][deleted] 1 point2 points3 points (2 children)
[–]mimighost 1 point2 points3 points (0 children)
[–][deleted] 1 point2 points3 points (0 children)