Title
Data classification with radial basis function networks based on a novel kernel density estimation algorithm.
Abstract
This paper presents a novel learning algorithm for efficient construction of the radial basis function (RBF) networks that can deliver the same level of accuracy as the support vector machines (SVMs) in data classification applications. The proposed learning algorithm works by constructing one RBF subnetwork to approximate the probability density function of each class of objects in the training data set. With respect to algorithm design, the main distinction of the proposed learning algorithm is the novel kernel density estimation algorithm that features an average time complexity of O(n log n), where n is the number of samples in the training data set. One important advantage of the proposed learning algorithm, in comparison with the SVM, is that the proposed learning algorithm generally takes far less time to construct a data classifier with an optimized parameter setting. This feature is of significance for many contemporary applications, in particular, for those applications in which new objects are continuously added into an already large database. Another desirable feature of the proposed learning algorithm is that the RBF networks constructed are capable of carrying out data classification with more than two classes of objects in one single run. In other words, unlike with the SVM, there is no need to resort to mechanisms such as one-against-one or one-against-all for handling datasets with more than two classes of objects. The comparison with SVM is of particular interest, because it has been shown in a number of recent studies that SVM generally are able to deliver higher classification accuracy than the other existing data classification algorithms. As the proposed learning algorithm is instance-based, the data reduction issue is also addressed in this paper. One interesting observation in this regard is that, for all three data sets used in data reduction experiments, the number of training samples remaining after a naive data reduction mechanism is applied is quite close to the number of support vectors identified by the SVM software. This paper also compares the performance of the RBF networks constructed with the proposed learning algorithm and those constructed with a conventional cluster-based learning algorithm. The most interesting observation learned is that, with respect to data classification, the distributions of training samples near the boundaries between different classes of objects carry more crucial information than the distributions of samples in the inner parts of the clusters.
Year
DOI
Venue
2005
10.1109/TNN.2004.836229
IEEE Transactions on Neural Networks
Keywords
Field
DocType
neural network,kernel density estimation,training sample,conventional cluster-based learning algorithm,algorithm work,radial basis function networks,radial basis function (rbf) network,data classification application,learning (artificial intelligence),novel kernel density estimation,kernel density estimation algorithm,data classification,data analysis,data reduction issue,rbf network,radial basis function network,support vector machine,data reduction experiment,machine learning,data classifier,learning algorithm,training data,algorithm design,probability density function,radial basis function,learning artificial intelligence,support vector,kernel density estimate,data reduction,time complexity
Data mining,One-class classification,Semi-supervised learning,Computer science,Wake-sleep algorithm,FSA-Red Algorithm,Artificial intelligence,Population-based incremental learning,Weighted Majority Algorithm,Online machine learning,Stability (learning theory),Pattern recognition,Algorithm,Machine learning
Journal
Volume
Issue
ISSN
16
1
1045-9227
Citations 
PageRank 
References 
79
2.95
22
Authors
5
Name
Order
Citations
PageRank
Yen-Jen Oyang142348.82
Shien-ching Hwang214110.55
Yu-Yen Ou325216.78
Chien-Yu Chen436729.24
Zhi-Wei Chen5792.95