Title
Towards Efficient Learning of Neural Network Ensembles from Arbitrarily Large Datasets
Abstract
Advances in data collection technologies allow accumulation of large and high dimensional datasets and provide opportunities for learning high quality classification and regression models. However, supervised learning from such data raises significant computational challenges including inability to preserve the data in computer main memory and the need for optimizing model parameters within given time constraints. For certain types of prediction models techniques have been developed for learning from large datasets, but few of them address efficient learning of neural networks. Towards this objective, in this study we proposed a procedure that automatically learns a series of neural networks of different complexities on smaller data chunks and then properly combines them into an ensemble predictor through averaging. Based on the idea of progressive sampling the proposed approach starts with a very simple network trained on a very small sample and then progressively increases the model complexity and the sample size until the learning performance no longer improves. Our empirical study on a synthetic and two real-life large datasets suggests that the proposed method is successful in learning complex concepts from large datasets with low computational effort.
Year
Venue
Keywords
2004
FRONTIERS IN ARTIFICIAL INTELLIGENCE AND APPLICATIONS
sample size,regression model,supervised learning,prediction model,empirical study,data collection,neural network
Field
DocType
Volume
Data mining,Competitive learning,Online machine learning,Semi-supervised learning,Computer science,Supervised learning,Unsupervised learning,Artificial intelligence,Artificial neural network,Ensemble learning,Arbitrarily large,Machine learning
Conference
110
ISSN
Citations 
PageRank 
0922-6389
3
0.46
References 
Authors
9
3
Name
Order
Citations
PageRank
Kang Peng117011.85
Zoran Obradovic21110137.41
Slobodan Vucetic363756.38