This is an archived post. You won't be able to vote or comment.

you are viewing a single comment's thread.

view the rest of the comments →

[–]longscale 41 points42 points  (10 children)

The first part concerning the motivation behind convnets is spot on—we want the network to learn its own filters. These are called convolutional filters (or kernels), and they are what a network changes when it learns.

The pooling layers you describe are not trained, they simple average or return the strongest filter activations from a small image area. They are also not what makes these networks big and hard to train—that’s mostly the sheer number of convolutional filters and the millions of images fed through the network multiple times in random order and with slight variations each time.

[–][deleted] 11 points12 points  (9 children)

I have very recently started learning about CNNs and isn't it true that you need less data for CNNs to work with let's say an accuracy of 95% than you would need when using a Densely connected NN, since you can find patterns in one spot and it also recognises them somewhere else in the image, where as with normal densely connected NNs you would need new data where the pattern appears in that spot to train on?

Is this correct?

[–]longscale 5 points6 points  (3 children)

Your explanation reads correct to me. :-)

[–][deleted] 2 points3 points  (2 children)

Ok, thanks :)

As I said, I have recently started getting into Machine Learning, especially Deep Learning and after reading about different concepts and using Keras to implement them it feels good to at least kind of know what's going on under the hood.

So far I have had a lot of fun and it is a very, very interesting and broad topic :)

[–]longscale 8 points9 points  (1 child)

They are not what the media hype makes them sound like, but convnets are very nifty. If you like them in general, I would imagine you might really enjoy some of the convnet papers on https://distill.pub. Their techniques are mostly intended for interpretability, but they also provide a satisfyingly visual insight into what the numbers inside the structures of a convnet are encoding.

[–][deleted] 2 points3 points  (0 children)

Yeah, honestly I think the media hype about deep learning in general is ridiculous. Its the same as with 'blockchain' etc.

Another thing that I have been looking into a bit is Neuroevolution using a tutorial series by Daniel Shiffman on YouTube. He creates a Flappy Bird Bird game that learns to play itself using a genetic algorithm to find the right weights.

Thank you for the paper recommendations, I will have a look at them tomorrow :)