Abstract | ||
---|---|---|
An ensemble is viewed as a machine learning system that combines multiple models to work collectively in the hope of producing a better performance than that of individuals. However, an ensemble's accuracy cannot be easily determined as it involves several factors, e.g. individual model's accuracy, diversity between its member models, decision-making strategy and number of members and the relationships between them are unclear. This paper, taking random decision tree ensembles as testing platforms, investigates these relationships and the strategies for creating ensembles from randomly generated trees. Specifically, we devised three sets of procedures for conducting experiments using twelve data sets from the UCI repository to determine the importance of individual model accuracy and the diversity between decision tree models within an ensemble. The main findings of the investigations are presented and discussed in the paper. |
Year | DOI | Venue |
---|---|---|
2006 | 10.1109/IJCNN.2006.247244 | 2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10 |
Keywords | Field | DocType |
machine learning,testing,voting,decision trees,artificial neural networks,bagging,training data,decision tree | Decision tree,Data mining,Computer science,Artificial intelligence,ID3 algorithm,Random forest,Ensemble learning,Alternating decision tree,Decision stump,Pattern recognition,Decision tree learning,Machine learning,Incremental decision tree | Conference |
ISSN | Citations | PageRank |
2161-4393 | 0 | 0.34 |
References | Authors | |
6 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Graeme Richards | 1 | 74 | 6.55 |
Wenjia Wang | 2 | 57 | 9.12 |