Omni-Sparsity DNN: Fast Sparsity Optimization for On-Device Streaming E2E ASR Via Supernet. | 0 | 0.34 | 2022 |
Collaborative Training of Acoustic Encoders for Speech Recognition. | 0 | 0.34 | 2021 |
Learning Dynamic Network Using a Reuse Gate Function in Semi-supervised Video Object Segmentation | 0 | 0.34 | 2021 |
MEMORY-EFFICIENT SPEECH RECOGNITION ON SMART DEVICES | 0 | 0.34 | 2021 |
Mixed Precision Training. | 0 | 0.34 | 2018 |
Mixed Precision Training. | 0 | 0.34 | 2018 |
Can FPGAs Beat GPUs in Accelerating Next-Generation Deep Neural Networks? | 60 | 2.55 | 2017 |
Hardware Accelerator For Analytics Of Sparse Data | 1 | 0.48 | 2016 |
Accelerating Binarized Neural Networks: Comparison of FPGA, CPU, GPU, and ASIC | 11 | 0.67 | 2016 |
Runnemede: An architecture for Ubiquitous High-Performance Computing | 38 | 1.17 | 2013 |
APE: accelerator processor extensions to optimize data-compute co-location | 0 | 0.34 | 2013 |
The GreenDroid Mobile Application Processor: An Architecture for Silicon's Dark Future | 76 | 3.42 | 2011 |
Efficient complex operators for irregular codes | 18 | 2.95 | 2011 |
An Evaluation of Selective Depipelining for FPGA-Based Energy-Reducing Irregular Code Coprocessors | 1 | 0.36 | 2011 |
Reducing the Energy Cost of Irregular Code Bases in Soft Processor Systems | 3 | 0.41 | 2011 |
QsCores: trading dark silicon for scalable energy efficiency with quasi-specific cores | 64 | 2.77 | 2011 |
Exploiting a computation reuse cache to reduce energy in network processors | 2 | 0.84 | 2005 |