Title
Comparison of Kullback-Leibler divergence approximation methods between Gaussian mixture models for satellite image retrieval
Abstract
In many applications, such as image retrieval and change detection, we need to assess the similarity of two statistical models. As a distance measure between two probability density functions, Kullback-Leibler divergence is widely used for comparing two statistical models. Unfortunately, for some models such as Gaussian Mixture Model (GMM), Kullback-Leibler divergence has no analytically tractable formula. We have to resort to approximation methods. In this paper, we compare seven methods, namely Monte Carlo method, matched bond approximation, product of Gaussian, variation-al method, unscented transformation, Gaussian approximation, and min-Gaussian approximation, for approximating the Kullback-Leibler divergence between two Gaussian mixture models for satellite image retrieval. Two image retrieval experiments based on two publicly available datasets have been performed. The comparison is carried out in terms of both retrieval performance and computational time.
Year
DOI
Venue
2015
10.1109/IGARSS.2015.7326631
IGARSS
Keywords
Field
DocType
Gaussian Mixture Model (GMM), Kullback-Leibler Divergence, Image Retrieval
Divergence-from-randomness model,Computer science,Image retrieval,Artificial intelligence,Gaussian function,Computer vision,Gaussian random field,Pattern recognition,Algorithm,Gaussian,Statistical model,Mixture model,Kullback–Leibler divergence
Conference
ISSN
Citations 
PageRank 
2153-6996
2
0.36
References 
Authors
5
2
Name
Order
Citations
PageRank
Shiyong Cui110311.54
Mihai Datcu2893111.62