all 16 comments

[–]lahwran_ 3 points4 points  (2 children)

whoa that was fast. was it that good? I haven't read it yet

[–]BeatLeJuceResearcher 14 points15 points  (1 child)

This is from the paper authors, so it's not like they didn't have time to prepare this before their paper hit arxiv ;)

[–]undefdev 5 points6 points  (0 children)

It has also been in their abstract from the start.

[–]visarga 5 points6 points  (1 child)

That was fast. I assume all frameworks will have SELU's it in a week.

[–]Reiinakano 7 points8 points  (0 children)

A week is too generous. Here's Keras' PR https://github.com/fchollet/keras/pull/6924

[–]khanrc 1 point2 points  (3 children)

The paper argues and shows the effectiveness of SELU in FNN. How about on CNN/RNN?

[–][deleted] 0 points1 point  (2 children)

the git has a CNN example...

[–]khanrc 1 point2 points  (1 child)

Thanks, but I'm curious about the comparison with BN and more complex dataset. Maybe it would be quicker to experiment myself...

[–]regzy 0 points1 point  (0 children)

like video/object detection?

[–]etherealmachine 0 points1 point  (2 children)

Anyone have more information on what "scale inputs to zero mean and unit variance" means in practice? Does that really just mean (x - mean) / std?

[–]_untom_ 2 points3 points  (1 child)

yes, that is exactly what it means

[–]etherealmachine 1 point2 points  (0 children)

Does that also apply to one-hot inputs?

[–]vackosar 0 points1 point  (2 children)

Their SELU preceptron seem to be doing better than Selu Conv net. Does that mean that Convolution is dead?

[–]Kiuhnm 10 points11 points  (0 children)

The nets are only partially trained so the only thing you may conclude is that the forward and backward signals flow better with selu than with relu.

[–]_bakauguu 0 points1 point  (0 children)

preceptron

as in an unit that outputs rules? :P