Title
Randomized Pruning: Efficiently Calculating Expectations in Large Dynamic Programs.
Abstract
Pruning can massively accelerate the computation of feature expectations in large models. However, any single pruning mask will introduce bias. We present a novel approach which employs a randomized sequence of pruning masks. Formally, we apply auxiliary variable MCMC sampling to generate this sequence of masks, thereby gaining theoretical guarantees about convergence. Because each mask is generally able to skip large portions of an underlying dynamic program, our approach is particularly compelling for high-degree algorithms. Empirically, we demonstrate our method on bilingual parsing, showing decreasing bias as more masks are incorporated, and outperforming fixed tic-tac-toe pruning.
Year
Venue
Field
2009
NIPS
Convergence (routing),Markov chain Monte Carlo,Computer science,Theoretical computer science,Artificial intelligence,Computation,Pruning,Principal variation search,Algorithm,Sampling (statistics),Pruning (decision trees),Parsing,Machine learning
DocType
Citations 
PageRank 
Conference
2
0.40
References 
Authors
20
3
Name
Order
Citations
PageRank
Bouchard-côté, Alexandre133218.57
Slav Petrov22405107.56
Dan Klein38083495.21