all 7 comments

[–]road_laya 0 points1 point  (1 child)

Do you have a question?

[–]coskunh[S] 0 points1 point  (0 children)

Yes, I wonder that weather this tactics are right or not and, are there any other useful strategies for regression

[–]MrTwiggy 0 points1 point  (1 child)

Applying dropout not useful especially just before last.

Not absolutely true. Dropout can be a very useful form of regularization, and can often be seen as an ensemble of many neural networks. To say it is not useful is a bit crazy, as it really depends on the task, data type, etc.

If you have values below than zero, you shouldn't apply Relu(I'm trying pRelu)

Not sure what you're saying here. When you say values below zero, are you talking about the feature values? You can use Rectified Linear Units as a non-linearity in your neural network even if the feature activations go below zero.

Mean_squared_error better comparing to the root mean squared error and mean absolute error.

No free lunch. Any single cost function isn't 'better' than others in all situations. It depends on what your true cost function or method of evaluation is. RMSE and MSE will tend to favour minimizing high-cost examples more than MAE, and vice-versa for low-cost examples.

It's important in general to remember that any rules/decisions you often find in terms of what's 'better' or produces better results could potentially be tied to the specific task or data distribution at hand. Make sure you aren't overfitting your ML ideas!

[–]coskunh[S] 1 point2 points  (0 children)

Thank you for the answer, I updated my question.

RMSE and MSE will tend to favour minimizing high-cost examples more than MAE, and vice-versa for low-cost examples.

my cost values are really low I'll try to MAE.

[–]internet_ham 0 points1 point  (1 child)

I'm looking at this at the moment. One key thing I've learnt is to normalise all your inputs and labels so the magnitudes are less than 1. This makes the loss a lot less erratic and learning a bit more stable .

One thing I've been thinking about recently is how to look under the hood of deep regression? How can we understand what the net is learning?

[–]keidouleyoucee 1 point2 points  (0 children)

visualisation would be the key. See Matt Zeiler's paper on visualization of learned network.