Recently, many l(1)-norm-based PCA approaches have been developed to improve the robustness of PCA. However, most existing approaches solve the optimal projection matrix by maximizing l(1)-norm-based variance and do not best minimize the reconstruction error, which is the true goal of PCA. Moreover, they do not have rotational invariance. To handle these problems, we propose a generalized robust metric learning for PCA, namely, l(2),(p)-PCA, which employs l(2),(p)-norm as the distance metric for reconstruction error. The proposed method not only is robust to outliers but also retains PCA's desirable properties. For example, the solutions are the principal eigenvectors of a robust covariance matrix and the low-dimensional representation have rotational invariance. These properties are not shared by l(1)-norm-based PCA methods. A new iteration algorithm is presented to solve l(2),(p)-PCA efficiently. Experimental results illustrate that the proposed method is more effective and robust than PCA, PCA-L1 greedy, PCA-L1 nongreedy, and HQ-PCA.
Page:
1336---1346
Related
Batch download
Cited By
noting
Similar Literature
Submit Feedback
Please wait while the file you selected is being converted