A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Towards Real-Time DNN Inference on Mobile Platforms with Model Pruning and Compiler Optimization
2020
Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence
High-end mobile platforms rapidly serve as primary computing devices for a wide range of Deep Neural Network (DNN) applications. However, the constrained computation and storage resources on these devices still pose significant challenges for real-time DNN inference executions. To address this problem, we propose a set of hardware-friendly structured model pruning and compiler optimization techniques to accelerate DNN executions on mobile devices. This demo shows that these optimizations can
doi:10.24963/ijcai.2020/751
dblp:conf/ijcai/BorghuisABF20
fatcat:nc5vj3jqinf6xku3upby5hnldy