Abstract | ||
---|---|---|
Can we learn a binary classifier from only positive data, without any negative data or unlabeled data? We show that if one can equip positive data with confidence (positive-confidence), one can successfully learn a binary classifier, which we name positive-confidence (Pconf) classification. Our work is related to one-class classification which is aimed at "describing" the positive class by clustering-related methods, but one-class classification does not have the ability to tune hyper-parameters and their aim is not on "discriminating" positive and negative classes. For the Pconf classification problem, we provide a simple empirical risk minimization framework that is model-independent and optimization-independent. We theoretically establish the consistency and an estimation error bound, and demonstrate the usefulness of the proposed method for training deep neural networks through experiments. |
Year | Venue | Keywords |
---|---|---|
2018 | ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018) | binary classification,binary classifier,supervised learning,positive data,classification problem,deep neural networks,one-class classification,proposed method |
DocType | Volume | ISSN |
Conference | 31 | 1049-5258 |
Citations | PageRank | References |
1 | 0.35 | 17 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Takashi Ishida | 1 | 12 | 5.23 |
Gang Niu | 2 | 204 | 36.78 |
Masashi Sugiyama | 3 | 3353 | 264.24 |