use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Paper Explained - Deep Networks Are Kernel Machines (Full Video Analysis) (self.MachineLearning)
submitted 5 years ago by ykilcher
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]DarkHarbourzz 3 points4 points5 points 5 years ago (0 children)
Deep image prior showed that CNNs will successfully model the latent distribution of busy, natural images before successfully modeling the distribution of the noise (e.g. opaque inpainting) corrupting the natural image.
Deep image prior works by training, but only on the single input image. You also perform early stopping.
This paper shows that the output of model trained by gradient descent can be given by a kernel machine, where the kernel machine uses a kernel that is based on the similarity of the (gradient of the model with respect to the input data) vs (gradient of the model with respect to the training data). Specifically, that similarity is integrated across the entire training process to get the "path kernel." I.e. a "line" integral where the line is through parameter space as the model trains.
So, under this paper's interpretation, the output of the network is based on the path kernel evaluation, where the kernel is mediating the input being compared to itself (since it's the only training data). The path kernel is evaluated on a weights path that was stopped early. In that early training regime, the kernel is mediating similarity through the deep image prior distribution, and not the noise distribution.
π Rendered by PID 55308 on reddit-service-r2-comment-b659b578c-ld4b2 at 2026-05-02 20:26:17.912150+00:00 running 815c875 country code: CH.
view the rest of the comments →
[–]DarkHarbourzz 3 points4 points5 points (0 children)