Feature scaling a preprocessing technique that is used to standardize the range of values in data features, making sure that the features are on a similar scale. It is used when the range of values of a certain feature is too variable and contains extreme values as most algorithms perform poorly when subjected to raw data which hasn’t been processed. Feature scaling can also allow algorithms such as gradient descent to operate much faster.
Feature scaling typically involves dividing every value in a data set by the range of values (maximum value - minimum value), resulting in a new range of just 1.
Feature scaling is often combined with mean normalization, which subtracts every value in a data set by the mean, resulting in a new mean of 0.
Combining feature scaling and mean normalization, we achieve a new feature z:
- z = (x - μ) / s
where μ is the mean of the value, s is the range of values, and x is the original feature.
- 1. NG A. "Machine Learning | Coursera". Coursera, 2018. [Link].
Related Radiopaedia articles
- artificial intelligence (AI)
- imaging data sets
- computer-aided diagnosis (CAD)
- natural language processing
machine learning (overview)
- machine learning processes
- machine learning models
- visualizing and understanding neural networks
- common data preparation/preprocessing steps
- DICOM to bitmap conversion
- dimensionality reduction
- principal component analysis
- training, testing and validation datasets
- loss function
- optimization algorithms
- linear and quadratic
- batch normalization
- rule-based expert systems