190 likes | 234 Views
Explore Dimensionality Reduction methods like PCA, Kernel PCA, and ICA for efficient data projection, signal representation, and classification. Learn through theory, algorithms, and examples in image and audio analysis. Discover optimal solutions and comparative analyses.
E N D
Feature Extraction (I) Data Mining IIYear 2009-10Lluís Belanche Alfredo Vellido
Principal Components Analysis (PCA) • General goal : project the data onto a new subspace so that a maximum of relevantinformation is preserved • In PCA, relevantinformation is variance (dispersion).
Two solutions: in which sense are they optimal? • In the signal representation sense • In the signal separation sense • In both • In none
Other approaches to FE • Kernel PCA: perform PCA in xΦ(x), where K(x,y) = < Φ(x), Φ(y)> is a kernel • ICA (Independent Components Analysis): • Seeks statistical independence of features (PCA seeks uncorrelated features) • Equivalence to PCA iff features are Gaussian • Image and audio analysis brings own methods • Series expansion descriptors (from the DFT, DCT or DST) • Moment-based features • Spectral features • Wavelet descriptors • Cao, J.J. et al. A comparison of PCA, KPCA and ICA for dimensionality reduction. Neurocomputing 55, pp. 321-336 (2003)