Title
Entropy Samplers and Strong Generic Lower Bounds For Space Bounded Learning.
Abstract
With any hypothesis class one can associate a bipartite graph whose vertices are the hypotheses H on one side and all possible labeled examples X on the other side, and an hypothesis is connected to all the labeled examples that are consistent with it. We call this graph the hypotheses graph. We prove that any hypothesis class whose hypotheses graph is mixing cannot be learned using less than Omega(log^2 |H|) memory bits unless the learner uses at least a large number |H|^Omega(1) labeled examples. Our work builds on a combinatorial framework that we suggested in a previous work for proving lower bounds on space bounded learning. The strong lower bound is obtained by defining a new notion of pseudorandomness, the entropy sampler. Raz obtained a similar result using different ideas.
Year
Venue
Field
2018
ITCS
Graph,Combinatorics,Vertex (geometry),Pseudorandomness,Upper and lower bounds,Bipartite graph,Omega,Mathematics,Bounded function
DocType
Citations 
PageRank 
Conference
2
0.39
References 
Authors
0
2
Name
Order
Citations
PageRank
Dana Moshkovitz136819.14
Michal Moshkovitz295.35