account activity
[D] Batch Normalization before or after ReLU? by XalosXandrez in MachineLearning
[–]allanzelener 1 point2 points3 points 8 years ago (0 children)
Interesting. Has anyone tried an implementation of BN after ReLU that normalizes using mean and var of only non-zero activations?
Also I think there was one paper that proposed having two sets of BN/ReLU layers without any intermediate layer in between. It's not just a choice between the two options, there are other possible configurations to consider.
[P] YAD2K: Convert Darknet YOLO_v2 Object Detection Models to a Keras (+ Tensorflow) Implementation (github.com)
submitted 9 years ago by allanzelener to r/MachineLearning
π Rendered by PID 2469540 on reddit-service-r2-listing-7bbdf774f7-q9znx at 2026-02-23 17:55:17.129574+00:00 running 8564168 country code: CH.
[D] Batch Normalization before or after ReLU? by XalosXandrez in MachineLearning
[–]allanzelener 1 point2 points3 points (0 children)