Abstract | ||
---|---|---|
In statistical pattern recognition, feature transformation attempts to change original feature space to a low-dimensional subspace, in which new created features are discriminative and non-redundant, thus improving the predictive power and generalization ability of subsequent classification models. Traditional transformation methods are not designed specifically for tackling data containing unreliable and noisy input features. To deal with these inputs, a new approach based on Dempster-Shafer Theory is proposed in this paper. A specific loss function is constructed to learn the transformation matrix, in which a sparsity term is included to realize joint feature selection during transformation, so as to limit the influence of unreliable input features on the output low-dimensional subspace. The proposed method has been evaluated by several synthetic and real datasets, showing good performance. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1007/978-3-319-40596-4_22 | Communications in Computer and Information Science |
Keywords | Field | DocType |
Belief functions,Dempster-Shafer theory,Feature transformation,Feature selection,Pattern classification | Feature transformation,Feature vector,Pattern recognition,Feature selection,Predictive power,Subspace topology,Computer science,Artificial intelligence,Transformation matrix,Discriminative model,Dempster–Shafer theory | Conference |
Volume | ISSN | Citations |
610 | 1865-0929 | 3 |
PageRank | References | Authors |
0.39 | 8 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Chunfeng Lian | 1 | 132 | 22.61 |
Ruan Su | 2 | 559 | 53.00 |
Thierry Denoeux | 3 | 815 | 74.98 |