You Already Have It: A Generator-Free Low-Precision DNN Training Framework Using Stochastic Rounding. | 0 | 0.34 | 2022 |
Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI | 2 | 0.36 | 2021 |
A Compression-Compilation Framework for On-mobile Real-time BERT Applications. | 0 | 0.34 | 2021 |
MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge. | 0 | 0.34 | 2021 |
Accelerating Framework of Transformer by Hardware Design and Model Compression Co-Optimization | 0 | 0.34 | 2021 |
NPAS: A Compiler-aware Framework of Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration | 1 | 0.35 | 2021 |
HMC-TRAN: A Tensor-core Inspired Hierarchical Model Compression for Transformer-based DNNs on GPU | 1 | 0.37 | 2021 |
Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning | 0 | 0.34 | 2020 |