Abstract | ||
---|---|---|
We consider semi-supervised learning, learning task from both labeled and unlabeled instances and in particular, self-training with decision tree learners as base learners. We show that standard decision tree learning as the base learner cannot be effective in a self-training algorithm to semi-supervised learning. The main reason is that the basic decision tree learner does not produce reliable probability estimation to its predictions. Therefore, it cannot be a proper selection criterion in self-training. We consider the effect of several modifications to the basic decision tree learner that produce better probability estimation than using the distributions at the leaves of the tree. We show that these modifications do not produce better performance when used on the labeled data only, but they do benefit more from the unlabeled data in self-training. The modifications that we consider are Naive Bayes Tree, a combination of No-pruning and Laplace correction, grafting, and using a distance-based measure. We then extend this improvement to algorithms for ensembles of decision trees and we show that the ensemble learner gives an extra improvement over the adapted decision tree learners. |
Year | DOI | Venue |
---|---|---|
2017 | 10.1007/s13042-015-0328-7 | Int. J. Machine Learning & Cybernetics |
Keywords | Field | DocType |
Semi-supervised learning, Self-training, Ensemble learning, Decision tree learning | Decision tree,Semi-supervised learning,Computer science,Artificial intelligence,ID3 algorithm,Ensemble learning,Machine learning,Decision tree learning,Alternating decision tree,Decision stump,Incremental decision tree | Journal |
Volume | Issue | ISSN |
8 | 1 | 1868-808X |
Citations | PageRank | References |
22 | 0.66 | 36 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jafar Tanha | 1 | 50 | 4.17 |
Maarten van Someren | 2 | 402 | 48.51 |
Hamideh Afsarmanesh | 3 | 1890 | 291.69 |