Title
Self-representation and PCA embedding for unsupervised feature selection.
Abstract
Feature selection is an important preprocessing step for dealing with high dimensional data. In this paper, we propose a novel unsupervised feature selection method by embedding a subspace learning regularization (i.e., principal component analysis (PCA)) into the sparse feature selection framework. Specifically, we select informative features via the sparse learning framework and consider preserving the principal components (i.e., the maximal variance) of the data at the same time, such that improving the interpretable ability of the feature selection model. Furthermore, we propose an effective optimization algorithm to solve the proposed objective function which can achieve stable optimal result with fast convergence. By comparing with five state-of-the-art unsupervised feature selection methods on six benchmark and real-world datasets, our proposed method achieved the best result in terms of classification performance.
Year
DOI
Venue
2018
10.1007/s11280-017-0497-2
World Wide Web
Keywords
Field
DocType
Feature selection, Subspace learning, Dimensionality reduction, Sparse learning, Principal component analysis
Clustering high-dimensional data,Dimensionality reduction,Embedding,Feature selection,Subspace topology,Pattern recognition,Computer science,Preprocessor,Regularization (mathematics),Artificial intelligence,Machine learning,Principal component analysis
Journal
Volume
Issue
ISSN
21
6
1386-145X
Citations 
PageRank 
References 
2
0.36
29
Authors
5
Name
Order
Citations
PageRank
Yonghua Zhu121612.38
Xuejun Zhang27016.55
Ruili Wang344650.35
Wei Zheng420832.78
Yingying Zhu5102.51