Technique in machine learning and statistics for dimensionality reduction and feature extraction.
1. Dimensionality Reduction:
2. Principal Components:
3. Orthogonality:
E.g. We have a hyperspectral dataset composed of 100 observations (pixels) in which each observation consists of 150 spectral features.
This dataset has too many features which mean a big dimensionality. If we have more features than observations, we have the risk of overfitting in supervised learning, so the performance will drop. Moreover, the observations become harder to cluster.
PCA can be solution for dimensionality problem. It projects the data along the directions where there is the largest variation of data. In this case by reducing the dimension to 10, we can reach a good balance between the number of observations and the number of features: 100/10 = 10.
1. Standardization:
2. Covariance Matrix Computation:
3. Eigen Decomposition:
4. Sort and Select:
5. Transform Data: