Model Compression Algorithm via Reinforcement Learning and Knowledge Distillation
-
Published:2023-11-09
Issue:22
Volume:11
Page:4589
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Liu Botao1, Hu Bing-Bing1, Zhao Ming1ORCID, Peng Sheng-Lung2ORCID, Chang Jou-Ming3ORCID
Affiliation:
1. School of Computer Science, Yangtze University, Jingzhou 434025, China 2. Department of Creative Technologies and Product Design, National Taipei University of Business, Taipei 10051, Taiwan 3. Institute of Information and Decision Sciences, National Taipei University of Business, Taipei 10051, Taiwan
Abstract
Traditional model compression techniques are dependent on handcrafted features and require domain experts, with a tradeoff between model size, speed, and accuracy. This study proposes a new approach toward resolving model compression problems. Our approach combines reinforcement-learning-based automated pruning and knowledge distillation to improve the pruning of unimportant network layers and the efficiency of the compression process. We introduce a new state quantity that controls the size of the reward and an attention mechanism that reinforces useful features and attenuates useless features to enhance the effects of other features. The experimental results show that the proposed model is superior to other advanced pruning methods in terms of the computation time and accuracy on CIFAR-100 and ImageNet dataset, where the accuracy is approximately 3% higher than that of similar methods with shorter computation times.
Funder
New Generation Information Technology Innovation Project
Subject
General Mathematics,Engineering (miscellaneous),Computer Science (miscellaneous)
Reference39 articles.
1. Krizhevsky, A., Sutskever, I., and Hinton, G.E. (2012, January 3–8). Imagenet classification with deep convolutional neural networks. Proceedings of the Advances in Neural Information Processing Systems 25, South Lake Tahoe, CA, USA. 2. Kortylewski, A., He, J., Liu, Q., and Yuille, A.L. (2020, January 13–19). Compositional convolutional neural networks: A deep architecture with innate robustness to partial occlusion. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 3. Kim, I., Baek, W., and Kim, S. (2020, January 13–19). Spatially attentive output layer for image classification. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 4. Guo, C., Fan, B., Zhang, Q., Xiang, S., and Pan, C. (2020, January 13–19). Augfpn: Improving multi-scale feature learning for object detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 5. Ren, S., He, K., Girshick, R., and Sun, J. (2015, January 7–12). Faster r-cnn: Towards real-time object detection with region proposal networks. Proceedings of the Advances in Neural Information Processing Systems 28, Montreal, QC, Canada.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|