Abstract
Principal component analysis (PCA) is well recognized in dimensionality reduction, and kernel PCA (KPCA) has also been proposed in statistical data analysis. However, KPCA fails to detect the nonlinear structure of data well when outliers exist. To reduce this problem, this paper presents a novel algorithm, named iterative robust KPCA (IRKPCA). IRKPCA works well in dealing with outliers, and can be carried out in an iterative manner, which makes it suitable to process incremental input data. As in the traditional robust PCA (RPCA), a binary field is employed for characterizing the outlier process, and the optimization problem is formulated as maximizing marginal distribution of a Gibbs distribution. In this paper, this optimization problem is solved by stochastic gradient descent techniques. In IRKPCA, the outlier process is in a high-dimensional feature space, and therefore kernel trick is used. IRKPCA can be regarded as a kernelized version of RPCA and a robust form of kernel Hebbian algorithm. Experimental results on synthetic data demonstrate the effectiveness of IRKPCA.
Original language | English |
---|---|
Pages (from-to) | 956-968 |
Number of pages | 13 |
Journal | International Journal of Computer Mathematics |
Volume | 87 |
Issue number | 5 |
Early online date | 10 May 2010 |
DOIs | |
Publication status | Published - 2010 |
Keywords
- dimension reduction
- feature extraction
- iterative robust KPCA
- KPCA
- outliers
- PCA