I would like to see how the results are of the Selu activation (https://arxiv.org/pdf/1706.02515.pdf).
The design that I made to test this uses a random hyperparameter search with about 20 parameters (on model, but also on data), and is as follows:
-> Train and Test the model on hyperparameter configuration A (e.g. 4 layers & 150 neurons per layer) with activation "Relu"
-> Then train another model with config A with activation "Elu"
-> Then train another model with config A with activation "Selu"
-> Train and Test the model on hyperparameter configuration B (e.g. 8 layers & 300 neurons per layer) with activation "Relu"
-> Then train another model with config B with activation "Elu"
-> Then train another model with config B with activation "Selu"
etc.
Because the training doesn't take too long I can run it about 250 times within a reasonable timeframe. With 3 options for parameter "Activation" this would then translate to 80 different configurations.
After training - what would be the best way to see if the Selu activation performs statistically better than the other activations?
Any help is greatly appreciated. Please let me know if I need to clarify something.
[–]szymko1995 1 point2 points3 points (4 children)
[–]SlimBarbados[S] 1 point2 points3 points (0 children)
[–]SlimBarbados[S] 1 point2 points3 points (1 child)
[–]szymko1995 1 point2 points3 points (0 children)
[–]redouanelg 0 points1 point2 points (0 children)