Title | ||
---|---|---|
Comparison of Kullback-Leibler divergence approximation methods between Gaussian mixture models for satellite image retrieval |
Abstract | ||
---|---|---|
In many applications, such as image retrieval and change detection, we need to assess the similarity of two statistical models. As a distance measure between two probability density functions, Kullback-Leibler divergence is widely used for comparing two statistical models. Unfortunately, for some models such as Gaussian Mixture Model (GMM), Kullback-Leibler divergence has no analytically tractable formula. We have to resort to approximation methods. In this paper, we compare seven methods, namely Monte Carlo method, matched bond approximation, product of Gaussian, variation-al method, unscented transformation, Gaussian approximation, and min-Gaussian approximation, for approximating the Kullback-Leibler divergence between two Gaussian mixture models for satellite image retrieval. Two image retrieval experiments based on two publicly available datasets have been performed. The comparison is carried out in terms of both retrieval performance and computational time. |
Year | DOI | Venue |
---|---|---|
2015 | 10.1109/IGARSS.2015.7326631 | IGARSS |
Keywords | Field | DocType |
Gaussian Mixture Model (GMM), Kullback-Leibler Divergence, Image Retrieval | Divergence-from-randomness model,Computer science,Image retrieval,Artificial intelligence,Gaussian function,Computer vision,Gaussian random field,Pattern recognition,Algorithm,Gaussian,Statistical model,Mixture model,Kullback–Leibler divergence | Conference |
ISSN | Citations | PageRank |
2153-6996 | 2 | 0.36 |
References | Authors | |
5 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Shiyong Cui | 1 | 103 | 11.54 |
Mihai Datcu | 2 | 893 | 111.62 |