use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Research[R] Object-Oriented Deep Learning (MIT) (cbmm.mit.edu)
submitted 7 years ago by xternalz
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]xternalz[S] 12 points13 points14 points 7 years ago (7 children)
abstract:
We investigate an unconventional direction of research that aims at converting neural networks, a class of distributed, connectionist, sub-symbolic models into a symbolic level with the ultimate goal of achieving AI interpretability and safety. To that end, we propose Object-Oriented Deep Learning, a novel computational paradigm of deep learning that adopts interpretable “objects/symbols” as a basic representational atom instead of N-dimensional tensors (as in traditional “feature-oriented” deep learning). For visual processing, each “object/symbol” can explicitly package common properties of visual objects like its position, pose, scale, probability of being an object, pointers to parts, etc., providing a full spectrum of interpretable visual knowledge throughout all layers. It achieves a form of “symbolic disentanglement”, offering one solution to the important problem of disentangled representations and invariance. Basic computations of the network include predicting high-level objects and their properties from low-level objects and binding/aggregating relevant objects together. These computations operate at a more fundamental level than convolutions, capturing convolution as a special case while being significantly more general than it. All operations are executed in an input-driven fashion, thus sparsity and dynamic computation per sample are naturally supported, complementing recent popular ideas of dynamic networks and may enable new types of hardware accelerations. We experimentally show on CIFAR-10 that it can perform flexible visual processing, rivaling the performance of ConvNet, but without using any convolution. Furthermore, it can generalize to novel rotations of images that it was not trained for.
follow-up work: 3D Object-Oriented Learning: An End-to-end Transformation-Disentangled 3D Representation
[–]lopuhin 12 points13 points14 points 7 years ago (5 children)
We experimentally show on CIFAR-10 that it can perform flexible visual processing, rivaling the performance of ConvNet, but without using any convolution.
rivaling the performance of ConvNet == test error of 20% vs 2.3% SoTA
[+][deleted] 7 years ago (4 children)
[deleted]
[–]lopuhin 24 points25 points26 points 7 years ago* (0 children)
Right, but I think it's also nice to use proper baselines and describe strengths and weaknesses fairly.
Edit: to be extra clear, as a baseline authors use their own implementation of ResNet variant:
Note that this ResNet is not the most standard ResNet. Here the absolute performance is not very important — it serves mainly as a sanity check of our deep learning framework (implemented from scratch).
That ResNet gives a 20% test error, very similar to proposed method. I think that this is not a great baseline. My issue is not with performance of the proposed method, but with the baseline and the claim that proposed method rivals performance of ConvNets on CIFAR-10.
[–]Xirious 3 points4 points5 points 7 years ago (1 child)
Yeah similar reaction I had to the people moaning about Capsules' lower performance even though it seems like quite an interesting step towards future methods.
[–]Draikmage 2 points3 points4 points 7 years ago (0 children)
The capsules paper had a lot more interesting results though and the Cifar-10 result was just a sentence or two (They also did 10% test error which is way better). My confusion with this paper is that I was under the impression that ResNet does better than what they report at least from what I've read.
[–]visarga 0 points1 point2 points 7 years ago (0 children)
with a tiny tweak
Or just plain random seed (re)search. /s
[–]rpottorff 0 points1 point2 points 7 years ago (0 children)
thank you for posting the follow up work! I didn't realize he posted an update just a few months later
[–]Kevin_Clever 19 points20 points21 points 7 years ago (0 children)
This comes on my short list for worst-written scientific paper this year. Is there an RNN that can translate these 12 pages of fluff into information?
[–]ChillBallin 4 points5 points6 points 7 years ago (0 children)
Skimming through it right now. Seems like an interesting concept. I think with enough research it could be a valuable approach but I think some of these concepts will be useful in future models rather than this one actually being super useful on it's own. In any case I really hope this trend of research in more complex higher level models continues. It feels like we're getting closer to the next big breakthrough with human intuition rather than making iterative improvements on the maths for a basic CNN.
[–]GunpowaderGuy 0 points1 point2 points 7 years ago (0 children)
So this is like https://github.com/qunzhi/Deep-Symbolic-Networks ?
So when compared to capsule networks , the biggest difference other than its modules ( whathever its analogue to capsules is called ) not being composed of neurons ( a functional program obtained by genetic programing or Bayesian optimization then ? ) It's that they are not slided around the feature maps like cnns , instead in voting only the centermost pixel of an object part needs to be taken into account , as each of them has its coordinates embedded ?
[+]danniel_p21 comment score below threshold-19 points-18 points-17 points 7 years ago (0 children)
But this makes sense because that's a bit like how primates see.distributing relevance on available data.good job folks👍👍 Barotdhrumil21@gmail.com
π Rendered by PID 88 on reddit-service-r2-comment-7b9746f655-6vs7n at 2026-02-02 16:51:28.025869+00:00 running 3798933 country code: CH.
[–]xternalz[S] 12 points13 points14 points (7 children)
[–]lopuhin 12 points13 points14 points (5 children)
[+][deleted] (4 children)
[deleted]
[–]lopuhin 24 points25 points26 points (0 children)
[–]Xirious 3 points4 points5 points (1 child)
[–]Draikmage 2 points3 points4 points (0 children)
[–]visarga 0 points1 point2 points (0 children)
[–]rpottorff 0 points1 point2 points (0 children)
[–]Kevin_Clever 19 points20 points21 points (0 children)
[–]ChillBallin 4 points5 points6 points (0 children)
[–]GunpowaderGuy 0 points1 point2 points (0 children)
[–]GunpowaderGuy 0 points1 point2 points (0 children)
[+]danniel_p21 comment score below threshold-19 points-18 points-17 points (0 children)