A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
SparseNN: An Energy-Efficient Neural Network Accelerator Exploiting Input and Output Sparsity
[article]
2017
arXiv
pre-print
Contemporary Deep Neural Network (DNN) contains millions of synaptic connections with tens to hundreds of layers. The large computation and memory requirements pose a challenge to the hardware design. In this work, we leverage the intrinsic activation sparsity of DNN to substantially reduce the execution cycles and the energy consumption. An end-to-end training algorithm is proposed to develop a lightweight run-time predictor for the output activation sparsity on the fly. From our experimental
arXiv:1711.01263v1
fatcat:ttorqjphzzg5xmrcasorbevbw4