Title
Discriminative Clustering: Optimal Contingency Tables by Learning Metrics
Abstract
The learning metrics principle describes a way to derive metrics to the data space from paired data. Variation of the primary data is assumed relevant only to the extent it causes changes in the auxiliary data. Discriminative clustering finds clusters of primary data that are homogeneous in the auxiliary data. In this paper, discriminative clustering using a mutual information criterion is shown to be asymptotically equivalent to vector quantization in learning metrics. We also present a new, finite-data variant of discriminative clustering and show that it builds contingency tables that detect optimally statistical dependency between the clusters and the auxiliary data. A finite-data algorithm is demonstrated to outperform the older mutual information maximizing variant.
Year
DOI
Venue
2002
10.1007/3-540-36755-1_35
ECML
Keywords
Field
DocType
auxiliary data,discriminative clustering,finite-data variant,older mutual information,metrics principle,optimal contingency tables,mutual information criterion,primary data,finite-data algorithm,data space,asymptotically equivalent,learning metrics,contingency table,mutual information
Fuzzy clustering,Correlation clustering,Pattern recognition,Vector quantization,Contingency table,Fisher information,Mutual information,Artificial intelligence,Paired Data,Cluster analysis,Mathematics
Conference
Volume
ISSN
ISBN
2430
0302-9743
3-540-44036-4
Citations 
PageRank 
References 
22
1.75
6
Authors
3
Name
Order
Citations
PageRank
Janne Sinkkonen123121.36
Samuel Kaski22755245.52
Janne Nikkilä320016.65