Abstract | ||
---|---|---|
Feature selection is an important preprocessing step for dealing with high dimensional data. In this paper, we propose a novel unsupervised feature selection method by embedding a subspace learning regularization (i.e., principal component analysis (PCA)) into the sparse feature selection framework. Specifically, we select informative features via the sparse learning framework and consider preserving the principal components (i.e., the maximal variance) of the data at the same time, such that improving the interpretable ability of the feature selection model. Furthermore, we propose an effective optimization algorithm to solve the proposed objective function which can achieve stable optimal result with fast convergence. By comparing with five state-of-the-art unsupervised feature selection methods on six benchmark and real-world datasets, our proposed method achieved the best result in terms of classification performance. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1007/s11280-017-0497-2 | World Wide Web |
Keywords | Field | DocType |
Feature selection, Subspace learning, Dimensionality reduction, Sparse learning, Principal component analysis | Clustering high-dimensional data,Dimensionality reduction,Embedding,Feature selection,Subspace topology,Pattern recognition,Computer science,Preprocessor,Regularization (mathematics),Artificial intelligence,Machine learning,Principal component analysis | Journal |
Volume | Issue | ISSN |
21 | 6 | 1386-145X |
Citations | PageRank | References |
2 | 0.36 | 29 |
Authors | ||
5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yonghua Zhu | 1 | 216 | 12.38 |
Xuejun Zhang | 2 | 70 | 16.55 |
Ruili Wang | 3 | 446 | 50.35 |
Wei Zheng | 4 | 208 | 32.78 |
Yingying Zhu | 5 | 10 | 2.51 |