Title
The Incredible Shrinking Neural Network: New Perspectives on Learning Representations Through The Lens of Pruning.
Abstract
How much can pruning algorithms teach us about the fundamentals of learning representations in neural networks? A lot, it turns out. Neural network model compression has become a topic of great interest in recent years, and many different techniques have been proposed to address this problem. In general, this is motivated by the idea that smaller models typically lead to better generalization. At the same time, the decision of what to prune and when to prune necessarily forces us to confront our assumptions about how neural networks actually learn to represent patterns in data. In this work we set out to test several long-held hypotheses about neural network learning representations and numerical approaches to pruning. To accomplish this we first reviewed the historical literature and derived a novel algorithm to prune whole neurons (as opposed to the traditional method of pruning weights) from optimally trained networks using a second-order Taylor method. We then set about testing the performance of our algorithm and analyzing the quality of the decisions it made. As a baseline for comparison we used a first-order Taylor method based on the Skeletonization algorithm and an exhaustive brute-force serial pruning algorithm. Our proposed algorithm worked well compared to a first-order method, but not nearly as well as the brute-force method. Our error analysis led us to question the validity of many widely-held assumptions behind pruning algorithms in general and the trade-offs we often make in the interest of reducing computational complexity. We discovered that there is a straightforward way, however expensive, to serially prune 40-70% of the neurons in a trained network with minimal effect on the learning representation and without any re-training.
Year
Venue
Field
2017
arXiv: Neural and Evolutionary Computing
Neural network learning,Computer science,Skeletonization,Through-the-lens metering,Artificial intelligence,Pruning (decision trees),Deep learning,Artificial neural network,Machine learning,Pruning,Computational complexity theory
DocType
Volume
Citations 
Journal
abs/1701.04465
3
PageRank 
References 
Authors
0.44
7
4
Name
Order
Citations
PageRank
Nikolas Wolfe130.44
Aditya Sharma230.44
Lukas Drude39511.10
Raj, Bhiksha42094204.63