Abstract
Weight pruning of deep neural networks (DNNs) has been proposed to satisfy the limited storage and computing capability of mobile edge devices. However, previous pruning methods mainly focus on reducing the model size and/or improving performance without considering the privacy of user data. To mitigate this concern, we propose a privacy-preserving-oriented pruning and mobile acceleration framework that does not require the private training dataset. At the algorithm level of the proposed framework, a systematic weight pruning technique based on the alternating direction method of multipliers (ADMM) is designed to iteratively solve the pattern-based pruning problem for each layer with randomly generated synthetic data. In addition, corresponding optimizations at the compiler level are leveraged for inference accelerations on devices. With the proposed framework, users could avoid the time-consuming pruning process for non-experts and directly benefit from compressed models. Experimental results show that the proposed framework outperforms three state-of-art end-to-end DNN frameworks, i.e., TensorFlow-Lite, TVM, and MNN, with speedup up to 4.2×, 2.5×, and 2.0×, respectively, with almost no accuracy loss, while preserving data privacy.
| Original language | English (US) |
|---|---|
| Title of host publication | GLSVLSI 2020 - Proceedings of the 2020 Great Lakes Symposium on VLSI |
| Publisher | Association for Computing Machinery |
| Pages | 119-124 |
| Number of pages | 6 |
| ISBN (Electronic) | 9781450379441 |
| DOIs | |
| State | Published - Sep 7 2020 |
| Externally published | Yes |
| Event | 30th Great Lakes Symposium on VLSI, GLSVLSI 2020 - Virtual, Online, China Duration: Sep 7 2020 → Sep 9 2020 |
Publication series
| Name | Proceedings of the ACM Great Lakes Symposium on VLSI, GLSVLSI |
|---|
Conference
| Conference | 30th Great Lakes Symposium on VLSI, GLSVLSI 2020 |
|---|---|
| Country/Territory | China |
| City | Virtual, Online |
| Period | 9/7/20 → 9/9/20 |
Bibliographical note
Publisher Copyright:© 2020 Association for Computing Machinery.
Keywords
- Model compression
- Pattern-based pruning
- Privacy-preserving
- Real-time mobile acceleration