Speaker
Can Hakan Dagidir
(KU Leuven)
Description
Kernel Principal Component Analysis (KPCA) extends linear PCA from a Euclidean space to data provided in the form of a kernel matrix. Several authors have studied its sensitivity to outlying cases and have proposed robust alternatives, as well as outlier detection diagnostics. We investigate the behavior of kernel ROBPCA, which relies on the Stahel-Donoho outlyingness in feature space (Debruyne and Verdonck, 2010). It turns out that one needs an appropriate selection of directions in which to compute the outlyingness. The outlier map of linear PCA also needs to be adapted to the kernel setting. Our study involves simulated and real data sets, such as the MNIST fashion data.
Type of presentation | Talk |
---|---|
Classification | Mainly methodology |
Keywords | robust statistics, kernel pca, outlier detection |