Z-score normalization: Difference between revisions

no edit summary
No edit summary
No edit summary
Line 1: Line 1:
{{see also|Machine learning terms}}
{{see also|Machine learning terms}}
==Introduction==
==Introduction==
Data normalization in machine learning is a critical preprocessing step that helps boost the performance of many algorithms. Normalization involves scaling data to a specified range or distribution to reduce the impact of differences in scale or units of features. One widely-used technique for normalization is Z-score normalization (also referred to as standardization).
Z-score normalization is a type of data scaling that transforms data values to have a mean of zero and standard deviation of one. This transformation occurs by subtracting the mean from each value and dividing by its standard deviation. The results are known as Z-scores, which indicate how far away from the mean each data point is.


==What is Z-score normalization?==
Data normalization in machine learning is a critical preprocessing step that helps boost the performance of many algorithms. Normalization involves scaling data to a specified range or distribution to reduce the impact of differences in scale or units of features.
Z-score normalization is a type of data scaling that transforms data values to have a mean of zero and standard deviation of one. This transformation occurs by subtracting the mean from each value and dividing by its standard deviation. The results are known as Z-scores, which indicate how far away from the mean each data point is.
 
==Example==
A feature with the mean of 500 and a standard deviation of 100
{| class="wikitable"
|
|-
! raw value
! Z-score
|-
| 500 || 0
|-
| 600 || 1
|-
| 355 || -1.45
|-
|}


==Why is Z-score normalization used?==
==Why is Z-score normalization used?==