Abstract | ||
---|---|---|
While unsupervised generative neural networks are attractive choices for adoption in always-on continuous-time smart sensory systems, they typically impose heavy memory requirements on the underlying computational fabric. Recent literature on binarized neural networks has not yet been extended to unsupervised generative networks and alternate strategies are required to reduce their memory footprint. This work studies unsupervised synaptic pruning strategies to reduce the memory requirements for Restricted Boltzmann Machines (RBMs). In addition to one-shot pruning, we explore alternative strategies that encompass iterative stochastic pruning as well as pruning under target probability density functions for an RBM trained over the MNIST database. Interestingly, the results presented here suggest that one-shot re-training after pruning of the least significant connections in a trained network yields improved performance/memory trade-off over multiple iterations of stochastic pruning and re-training on the same network. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1109/BIOCAS.2018.8584839 | 2018 IEEE BIOMEDICAL CIRCUITS AND SYSTEMS CONFERENCE (BIOCAS): ADVANCED SYSTEMS FOR ENHANCING HUMAN HEALTH |
Field | DocType | ISSN |
Computer vision,Boltzmann machine,MNIST database,Computer science,Memory management,Artificial intelligence,Artificial neural network,Memory footprint,Synaptic pruning,Probability density function,Machine learning,Pruning | Conference | 2163-4025 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Surabhi Kalyan | 1 | 0 | 0.34 |
Siddharth Joshi | 2 | 19 | 2.02 |
S Sheik | 3 | 30 | 5.20 |
Bruno Pedroni | 4 | 86 | 8.46 |
Gert Cauwcnbcrghs | 5 | 0 | 0.34 |