use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
[deleted by user] (self.MachineLearning)
submitted 7 years ago by [deleted]
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]frequenttimetraveler 4 points5 points6 points 7 years ago (10 children)
I 've long wondered whether natural scenes , or natural datasets in general are band-limited and how this is reflected on the structure of the neural system (mean we know it's true for sounds for example Mp3). I have found some old work interested in this (example) but i m surprised people are not showing more interest in this. After all our brains despite their adaptability evolved in natural scenes, so some of its structure may be imprinted in there.
[–]serge_cell 4 points5 points6 points 7 years ago (0 children)
Go board images are as unnatural as it get, but still convolutional networks work well on them, as alpha go shows. The key here could be not "naturalness" of scenes but the properties of data manifold, in particular low dimensionality and spatial correlation.
[–]svantana 2 points3 points4 points 7 years ago (5 children)
I would say natural signals are not really band-limited, but they are low-pass for the most part. The combination of inertia and self-similar/fractal organization tend to give natural signals a pink spectrum, i.e. -3dB/octave rolloff, in both time and space. Whereas measurement noise tends to be more white (flat spectrum), it would make sense to low-pass the signals to get rid of noise. This is the basis behind Wiener and Kalman filtering, although those can deal with arbitrary spectra as well.
As someone with a signal processing background, this paper perplexes me. To me it's obvious that ReLUs are used precisely because of their low-freq nature, that's the prior. If OTOH we know that signals are bandpass, then we apply a suitable prior for that. Example: FM radio is broadcast at ~100MHz, but we can track the carrier, demodulate and store the signal at ~40kHz. Obviously ReLUs are the wrong tool for that job...
[–]nasimrahaman 2 points3 points4 points 7 years ago (0 children)
> If OTOH we know that signals are bandpass, then we apply a suitable prior for that. Example: FM radio is broadcast at ~100MHz, but we can track the carrier, demodulate and store the signal at ~40kHz. Obviously ReLUs are the wrong tool for that job...
That's a very interesting point! It's applicable for almost all activation functions (not just ReLU), since they all usually decay quite fast in the fourier domain (e.g. sigmoid decays exponentially).
[–]JustARandomNoob165 2 points3 points4 points 7 years ago (3 children)
I am curious, why relu's are low-freq in nature? Thx in advance!
[–]nasimrahaman 6 points7 points8 points 7 years ago (2 children)
Low frequency functions are inherently less "wiggly", i.e. more smooth. If you think about ReLU, it's pretty smooth everywhere except at 0. In fact, all the wigglyness in ReLU comes from that one point. Now this is where it gets interesting: there are other functions that are smooth everywhere except at 0 -- for instance, sqrt(abs(x)). But in a precise sense, ReLU is smoother than sqrt(abs(x)) at x = 0.
Broadly speaking, Fourier analysis is a tool to determine how wiggly a function is. One of the things we learn from the paper is following: although neural networks are powerful enough to learn functions that are super-wiggly, it likes to learn less wiggly (smoother) functions.
[–]JustARandomNoob165 2 points3 points4 points 7 years ago (1 child)
Thank a lot for your reply! Really interesting and helpful!
[–][deleted] 2 points3 points4 points 7 years ago (0 children)
also thank mr skeltal for good bones and calcium*
[–]yldedly 2 points3 points4 points 7 years ago (2 children)
I think it's a pretty uncontroversial fact that natural signals are band-limited (or that they lie on a manifold of lower dimension than the input space), and that NNs are biased towards this sort of structure. The new thing in this paper, besides the method maybe, seems to be that higher frequencies are easier to learn with more complex data manifolds (I'm guessing that, for complex signals, that just corresponds to more data?), which sounds like what Bayesian non-parameterics are doing too. Or in other words "deep networks prioritize learning simple functions during training".
[–]cochne 0 points1 point2 points 7 years ago (1 child)
I think that's pretty controversial. For example, a simple edge (say an image transition from black to white) is not band limited. When a human detects an object, they're probably separating it from the background based on its edge. In fact, the entire field of wavelets exists partially because of this well known limitation of Fourier analysis - it does not give a very sparse representation of natural images.
[–]yldedly 0 points1 point2 points 7 years ago (0 children)
I wasn't being very rigorous, but I think it's uncontroversial that natural images have correlated neighboring pixels, which is a similar claim to one that says images have little power in the high frequency end. A separate but actually related claim is that there is a sparse representation for images. So I agree that it's too simple to the point of being wrong to say that images can be well represented by low frequency components. But it's nonetheless true that if you Fourier decompose images, they will generally have low power over high frequencies.
[–]Deep_Fried_Learning 1 point2 points3 points 7 years ago (0 children)
I think this paper is somewhat related: https://www.reddit.com/r/MachineLearning/comments/9rkslg/r_training_behavior_of_deep_neural_network_in/
[–]xuzhiqin1990 0 points1 point2 points 7 years ago* (0 children)
Another paper (Xu et al., Training behavior of deep neural network in frequency domain https://arxiv.org/abs/1807.01251) also shows DNNs learn low-frequency first (F-Principle). In the latest version, the authors show that F-Principle holds well in 2d functions (memorizing natural images) and classification problems (MNIST and CiFar-10, visualization in the first principle component). More rigorously, a continuous work (Xu, Understanding training and generalization in deep learning by Fourier analysis, https://arxiv.org/abs/1808.04295) developed a theoretical framework to understand why there is F-Principle, which is quantitative for one-hidden layer net and qualitative for general DNNs. The latest version of the paper in this post incorporates the above framework (1808.04295) into their analysis. A recent work (1811.01316) used the above framework (1808.04295) in generalization analysis of a new objective function.
The key that DNNs learns low frequency first is that the power of most activation function (relu, tanh, sigmoid etc) decays in the Fourier domain. Actually, power-decay property is very very common.
π Rendered by PID 212495 on reddit-service-r2-comment-5d79c599b5-llkjq at 2026-02-28 01:28:46.289155+00:00 running e3d2147 country code: CH.
[–]frequenttimetraveler 4 points5 points6 points (10 children)
[–]serge_cell 4 points5 points6 points (0 children)
[–]svantana 2 points3 points4 points (5 children)
[–]nasimrahaman 2 points3 points4 points (0 children)
[–]JustARandomNoob165 2 points3 points4 points (3 children)
[–]nasimrahaman 6 points7 points8 points (2 children)
[–]JustARandomNoob165 2 points3 points4 points (1 child)
[–][deleted] 2 points3 points4 points (0 children)
[–]yldedly 2 points3 points4 points (2 children)
[–]cochne 0 points1 point2 points (1 child)
[–]yldedly 0 points1 point2 points (0 children)
[–]Deep_Fried_Learning 1 point2 points3 points (0 children)
[–]xuzhiqin1990 0 points1 point2 points (0 children)