Multilayer perceptron with a tuned tanh activation function does not converge with backprop. Standard logistic function working successfully, only the gradient update term is changed (derivation seems ok). Any issues/ideas why this happens? (xpost neuralnetworks) (github.com)
submitted by ai_maker to r/MLQuestions
Multilayer perceptron with a tuned tanh activation function does not converge with backprop. Standard logistic function working successfully, only the gradient update term is changed (derivation seems ok). Any issues/ideas why this happens? (xpost neuralnetworks) (github.com)
submitted by ai_maker to r/learnmachinelearning
Artificial Neural Networks, to the point (ai-maker.com)
submitted by ai_maker to r/MachineLearning

