Abstract | ||
---|---|---|
Representation learning models, such as the sparse representation and low-rank representation, have shown pleasing efficacy in exploring the intrinsic data structures for patternrecognition tasks. However, conventional methods ignore the local geometric and similarity information among samples, and the performance is restricted. To address this issue, this paper proposes a novel Data Induced Masking Representation (DIMR) learning model by imposing explicit regularization and low-rank constraint. Specifically, DIMR is formulated for shrinking the representations of inter-class and non-neighbor samples. An extra representation regularization term is deployed with a data induced mask matrix, which can incorporate label and locality priors to guide the learning of affinity representation matrix. The affinity graph derived from DIMR is with low-rank, locality preservation (sparsity) and label guiding, such that it can better characterize the adjacent relationship between samples. Extensive experiments on benchmark face datasets demonstrate the superiority of DIMR for both semi-supervised classification and semi-supervised subspace learning tasks. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1016/j.knosys.2019.04.006 | Knowledge-Based Systems |
Keywords | Field | DocType |
Representation learning,Semi-supervised learning,Low-rank representation,Sparse representation,Subspace learning | Data structure,Locality,Pattern recognition,Subspace topology,Matrix (mathematics),Computer science,Sparse approximation,Regularization (mathematics),Artificial intelligence,Prior probability,Machine learning,Feature learning | Journal |
Volume | ISSN | Citations |
177 | 0950-7051 | 1 |
PageRank | References | Authors |
0.35 | 0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Tan Guo | 1 | 18 | 3.85 |
Le Zhang | 2 | 195 | 20.62 |
Xiao-heng Tan | 3 | 11 | 11.05 |
Liu Yang | 4 | 18 | 3.80 |
Zhifang Liang | 5 | 1 | 0.69 |