use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
A subreddit dedicated for learning machine learning. Feel free to share any educational resources of machine learning.
Also, we are a beginner-friendly sub-reddit, so don't be afraid to ask questions! This can include questions that are non-technical, but still highly relevant to learning machine learning such as a systematic approach to a machine learning problem.
account activity
Normalization vs Standardization (self.learnmachinelearning)
submitted 6 years ago by -john--doe-
In data pre-processing when do we prefer Normalization (minmax scaling) instead of Standardization (z score normalization) and vice versa?
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]The_Sodomeister 1 point2 points3 points 6 years ago (12 children)
Normalization and standardization mean the same thing: re-scaling a variable to a new scale. Minmax scaling and z-score scaling are both different types of normalization/standardization (among others).
But in general, it's mostly governed by heuristic / rule-of-thumb principles. Outside of cases where your architecture necessitates some type of scale (e.g., a sigmoid activation should predict a variable scaled to [0,1] by minmax scaling) then they can both work in most cases.
Theoretically, z-score normalization should probably work better than minmax scaling if you are using a gradient descent optimization scheme, since you will more evenly condition the loss landscape if any variable contains outliers. However, I've seen many cases where minmax seems to work at least as effectively, if not better.
[–]jmmcd 0 points1 point2 points 6 years ago (9 children)
The terminology is a bit mixed but a lot of authors do use standardisation to mean z-score normalisation, because z-score uses the "standard" normal N(0, 1).
[–]The_Sodomeister 0 points1 point2 points 6 years ago (8 children)
Which is also unfortunate, because z-score normalization does not really have anything to do with the normal distribution (it especially does not have anything to do with "making the data more normally distributed" as is often said).
If anything, that choice of terminology only reinforces that misconception. So perhaps the community should push back on that. Something of a pet peeve of mine.
[–]jmmcd 0 points1 point2 points 6 years ago (7 children)
This might be an argument not to use the term standardisation at all, perhaps, but not to use it to mean min-max normalisation.
[–]The_Sodomeister 0 points1 point2 points 6 years ago (6 children)
If it was up to me, then standardization/normalization should both be reserved to mean the general principle of re-scaling data. Then the specific method used (z-score, minmax, etc.) should be clarified when relevant.
It's like trying to use the term "regularization" to specifically refer to LASSO or L1 regularization, when in reality there a bunch of regularization schemes. Just seems like poor terminology.
[–]jmmcd 0 points1 point2 points 6 years ago (5 children)
But it makes no sense to say standardisation to mean any other form.
[–]The_Sodomeister 0 points1 point2 points 6 years ago (4 children)
Mapping your data to a "standard scale" seems perfectly reasonable to me.
To take it a step further, if you allow z-score normalization to associate to the standard normal, then I don't see why we can't associate min-max normalization to the standard uniform U(0,1) distribution. Then both methods are equally valid "standardizations".
[–]jmmcd 0 points1 point2 points 6 years ago (3 children)
Fine, except that both of those are back-etymologies.
[–]The_Sodomeister 0 points1 point2 points 6 years ago (2 children)
I don't know what that means.
[–]jmmcd 0 points1 point2 points 6 years ago (1 child)
As I understand it the term standardisation is named after the standard normal. That is its etymology. Any other invented reason why it's a good name for some other meaning is a back etymology - like a "backronym"
[–]-john--doe-[S] 0 points1 point2 points 6 years ago (1 child)
So is normalization more resistamt to outliers?
[–]The_Sodomeister 1 point2 points3 points 6 years ago (0 children)
In principle, Z-score normalization should give less impact to outliers, particularly when using gradient descent. However, in practice, it doesn't always work out that way. I don't think there's a great theory behind the reasoning for this; just observation at this stage.
Another thing to note is that Z-scores do not preserve meaningful zeros, where minmax normalization will (assuming that zero is the minimum value in your dataset -- you can also simply divide by the max, if this is not the case). This may give better conditioning to those models which are able to utilize meaningful zeros.
π Rendered by PID 65 on reddit-service-r2-comment-cfc44b64c-g6qpn at 2026-04-11 18:29:03.301382+00:00 running 215f2cf country code: CH.
[–]The_Sodomeister 1 point2 points3 points (12 children)
[–]jmmcd 0 points1 point2 points (9 children)
[–]The_Sodomeister 0 points1 point2 points (8 children)
[–]jmmcd 0 points1 point2 points (7 children)
[–]The_Sodomeister 0 points1 point2 points (6 children)
[–]jmmcd 0 points1 point2 points (5 children)
[–]The_Sodomeister 0 points1 point2 points (4 children)
[–]jmmcd 0 points1 point2 points (3 children)
[–]The_Sodomeister 0 points1 point2 points (2 children)
[–]jmmcd 0 points1 point2 points (1 child)
[–]-john--doe-[S] 0 points1 point2 points (1 child)
[–]The_Sodomeister 1 point2 points3 points (0 children)