Title
Minimum Description Length Codes Are Critical.
Abstract
In the Minimum Description Length (MDL) principle, learning from the data is equivalent to an optimal coding problem. We show that the codes that achieve optimal compression in MDL are critical in a very precise sense. First, when they are taken as generative models of samples, they generate samples with broad empirical distributions and with a high value of the relevance, defined as the entropy of the empirical frequencies. These results are derived for different statistical models (Dirichlet model, independent and pairwise dependent spin models, and restricted Boltzmann machines). Second, MDL codes sit precisely at a second order phase transition point where the symmetry between the sampled outcomes is spontaneously broken. The order parameter controlling the phase transition is the coding cost of the samples. The phase transition is a manifestation of the optimality of MDL codes, and it arises because codes that achieve a higher compression do not exist. These results suggest a clear interpretation of the widespread occurrence of statistical criticality as a characterization of samples which are maximally informative on the underlying generative process.
Year
DOI
Venue
2018
10.3390/e20100755
ENTROPY
Keywords
Field
DocType
Minimum Description Length,normalized maximum likelihood,statistical criticality,phase transitions,large deviations
Statistical physics,Pairwise comparison,Mathematical optimization,Boltzmann machine,Phase transition,Minimum description length,Coding (social sciences),Statistical model,Criticality,Dirichlet distribution,Mathematics
Journal
Volume
Issue
ISSN
20
10
1099-4300
Citations 
PageRank 
References 
0
0.34
0
Authors
3
Name
Order
Citations
PageRank
Ryan John Cubero100.34
Matteo Marsili214917.65
Yasser Roudi382.22