Title
Distribution based classification using Gaussian Mixture Models.
Abstract
A central task in classification is a measure of similarity between a dataset and a class that is characterised by a probability density function. The Bhattacharyya distance and the Kullback-Liebler divergence measure have been successful in comparing two multivariate normal density functions but their use is impracticable when the data is modelled using complex distributions such as Gaussian Mixture Models. The similarity is computed by combining the Bhattacharyya distances between corresponding mixtures in the reference and the test data model. In this paper we compare the performance of the Likelihood Ratio Test to a novel technique that defines a similarity measure between data and reference models having Gaussian Mixture probability density functions. When fitting a Gaussian Mixture Model to the test dataset our procedure ensures a one to one correspondence between the mixtures of the dataset and those of the reference model. This procedure has been tested using experiments, with both synthetic data and a Speaker Verification evaluation database. The performance was assessed using Detection Error Trade-off curves and demonstrates that the new measure performs significantly better than Likelihood Ratio Test.
Year
DOI
Venue
2002
10.1109/ICASSP.2002.5745576
ICASSP
Keywords
Field
DocType
multivariate normal,synthetic data,gaussian mixture model,reference model,probability density function,likelihood ratio test,data model,speech
Bhattacharyya distance,Similarity measure,Pattern recognition,Likelihood-ratio test,Gaussian,Multivariate normal distribution,Test data,Artificial intelligence,Probability density function,Mathematics,Mixture model
Conference
Volume
ISSN
ISBN
4
1520-6149
0-7803-7402-9
Citations 
PageRank 
References 
0
0.34
0
Authors
2
Name
Order
Citations
PageRank
Jon Gudnason1645.76
Mike Brookes210410.08