Generated by GPT-5-mini| PCA | |
|---|---|
| Name | Principal component analysis |
| Caption | Scatter plot with principal components |
| Classification | Dimensionality reduction |
| Input | Dataset matrix |
| Output | Principal components |
| Introduced | 1901 |
| Inventor | Karl Pearson |
| Field | Statistics |
PCA
Principal component analysis is a statistical technique for dimensionality reduction and feature extraction that transforms correlated variables into a smaller set of uncorrelated components. It was introduced by Karl Pearson and further developed by Harold Hotelling; PCA underpins methods in multivariate analysis, signal processing, and pattern recognition. PCA finds application across domains associated with Francis Galton-era biometric research, Andrey Kolmogorov-inspired stochastic modeling, and modern computational efforts led by institutions such as AT&T Bell Laboratories and MIT.
PCA originated in early 20th-century work by Karl Pearson and later theoretical framing by Harold Hotelling. It projects data onto orthogonal directions that maximize variance, linking to eigenanalysis introduced by David Hilbert and matrix algebra advanced by James Joseph Sylvester. PCA connects historically to empirical studies at University of Cambridge and to multivariate techniques disseminated through texts by R. A. Fisher and John Tukey. Practically, PCA is used in pipelines at organizations ranging from NASA to Google for tasks including compression, denoising, and exploratory analysis.
Given a centered data matrix X (observations by variables), PCA solves an eigenvalue problem for the covariance matrix C = (1/(n-1)) X^T X, relating to linear algebra work by Carl Friedrich Gauss and Augustin-Louis Cauchy. Principal components are eigenvectors of C associated with descending eigenvalues, echoing spectral theory developed by David Hilbert and Erhard Schmidt. Equivalent formulations use singular value decomposition (SVD) of X as in algorithms from Gene H. Golub and William Kahan; singular values squared correspond to explained variances linked to concepts formalized by Andrey Kolmogorov and Norbert Wiener in stochastic processes. Dimensionality reduction selects the first k components to minimize reconstruction error in a least-squares sense, relating to projection theory in the work of John von Neumann.
Classical computation uses eigen-decomposition of the covariance matrix via methods traced to John von Neumann and implemented in libraries from Netlib contributors such as Jim Demmel. SVD-based algorithms by Gene H. Golub and William Kahan provide numeric stability and are common in packages from LAPACK and BLAS ecosystems developed with contributions from Jack Dongarra. For large-scale data, iterative methods like the power method and Lanczos algorithm—credited to Cornelius Lanczos—and randomized algorithms pioneered by teams at Stanford University and IBM enable scalable PCA. Implementations appear in software by Microsoft Research, Google Research, and open-source projects like Scikit-learn and TensorFlow.
PCA is employed in image compression work associated with Yann LeCun-influenced deep learning research and in remote sensing analyses by European Space Agency teams. In genetics, PCA helps interpret population structure in studies by groups at The Broad Institute and Wellcome Trust Sanger Institute. PCA aids face recognition systems developed at MIT Media Lab and surveillance research tied to DARPA programs. In finance, analysts at institutions like Goldman Sachs and J.P. Morgan use PCA for risk factor modeling, while climatologists at NOAA and Hadley Centre apply PCA to identify dominant modes such as teleconnection patterns explored by Edward Lorenz. PCA also supports neuroscience research at Harvard Medical School and Max Planck Society labs for dimensionality reduction of neural recordings.
Interpreting components requires care: loadings can be influenced by scaling decisions linked to standardization practices recommended by statisticians like Karl Pearson and R. A. Fisher. PCA assumes linearity and maximizes variance, which can misrepresent structure in datasets relevant to nonlinear manifolds studied by Benoît Mandelbrot and Richard Hamming. The method is sensitive to outliers, a problem highlighted in robust statistics literature by Peter J. Huber and remedied by alternatives from Frank Hampel. PCA components lack guaranteed identifiability without constraints, echoing identifiability issues discussed by Jerzy Neyman and Egon Pearson. Suitability must be evaluated in contexts such as genomic stratification analyses at Wellcome Trust and image-processing pipelines at Bell Labs.
Extensions include kernel methods developed by Bernhard Schölkopf and Vladimir Vapnik—kernel PCA—that capture nonlinear structure; sparse PCA variants influenced by work at University of California, Berkeley and optimization contributions by Stephen Boyd; probabilistic PCA formulated by Michael I. Jordan and Zoubin Ghahramani; and independent component analysis advanced by Aapo Hyvärinen and Teuvo Kohonen for statistical independence criteria. Other adaptations include robust PCA frameworks motivated by Emmanuel Candès and Terence Tao for matrix decomposition, and multilinear/tensor PCA applied in studies at Rice University and University of Illinois Urbana-Champaign. Randomized PCA algorithms from researchers at Stanford University and MIT enhance scalability for big-data applications used by Facebook and Amazon.
Category:Statistical methods