270 likes | 475 Views
A principal components analysis self-organizing map. Ezequiel Lopez-Rubio, Jose Munoz-Perez, Jose Antonio Gomez-Ruiz. Advisor : Dr. Hsu Student : Sheng-Hsuan Wang Department of Information Management. Neural Network 17 (2004) 261-270. Outline. Motivation Objective
E N D
A principal components analysis self-organizing map Ezequiel Lopez-Rubio, Jose Munoz-Perez, Jose Antonio Gomez-Ruiz Advisor : Dr. Hsu Student : Sheng-Hsuan Wang Department of Information Management Neural Network 17 (2004) 261-270
Outline • Motivation • Objective • The ASSOM network • The PCASOM model • Experiments • Conclusion
Motivation • The adaptive subspace self-organizing map (ASSOM) is an alternative to the standard principal component analysis (PCA) algorithm • Look for the most relevant features of the input data. • However, its training equations are complexes. • Separate ability in the classical PCA and ASSOM.
Objective • This paper proposed a new self-organizing neural model that performs principal components analysis • Like the ASSOM, but has a broader capability to represent the input distribution.
The ASSOM network • The ASSOM network uses subspaces in each node rather than just single weights. • The ASSOM network is based on training not just using single samples but sets of slightly translated, rotated and/or scaled signal or image samples, called episodes. • Each neuron of an ASSOM network represents a subset of the input data with a vector basis which is adapted so that the local geometry of the input data is build.
The ASSOM network projection error orthogonal projection
The ASSOM network • orthogonal projection • A vector x on an orthonormal vector basis B={bh|h=1,…,K} • The vector x can be decomposed into two vectors • orthogonal projection and projection error.
The ASSOM network • The input vectors are grouped into episodes in order to be presented to the network. • An episode S(t) has many time instants tp belongs to S(t), each with an input vector x(tp). • episodes: sets of slightly translated, rotated or scaled samples.
The ASSOM network • Winner lookup • Learning • Basis vectors rotation • Dissipation Eliminate instability • Orthonormalization • Orthonormalize every basis for good performance.
The ASSOM network • The objective function is the average expected spatially weighted normalized squared projection error over the episodes. • The Robbins-Monro stochastic approximation is used to minimize objective function, which leads to Basis vectors rotation.
The PCASOM network • Neuron weights updating • Use the covariance matrix to store the information. • The covariance matrix of an input vector x is defined as • M input samples
The PCASOM network • The best approximation • It is an unbiased estimator with minimum variance. • If we obtain N new input samples
The PCASOM network The outputs of the algorithm are RV and eIII the new approximations.
The PCASOM network • Competition among neurons • The neuron c that has the minimum sum of projection errors is the winner: • Orth(x, B) is the orthogonal projection of vector x on basis B.
The PCASOM network • Network topology • Neighborhood function • update the vector ei and the matrix Ri
The PCASOM network • Summary • For every unit i, obtain the initial covariance matrixR(0). • For every unit i, build the vector ei(0) by using small random value. • At time instant t, select the input vectors x(t). Compute the winning neuron c. • For every unit i; update the vector ei and the matrix Ri • Convergence condition.
Comparison with ASSOM • Solidly rooted on statistics. • Update equation is more stable • Matrix sums • Does not need episodes. • It has a wider capability to represent the input distribution.
Experiments • Convergence speed experiment • Relative error for an input vector x projection error norm for BMU the norm of the input vector
Experiments • Separation capability experiment
Experiments • UCI benchmark databases experiment
Experiments • UCI benchmark databases experiment
Conclusions • A new self-organizing network that performs PCA • Related to the ASSOM • Its training equations are much simpler • Its input representation capability is broader • Experiments show that the new model has better performance than the ASSOM network.
Personal opinion • Valuable idea • SOM based on PCA • Contribution • Input data • Cluster shape • Performance • Drawback • Hard to implement.