Pruning Deep Neural Network Models via Minimax Concave Penalty Regression

Author:

Liu Xinggu1,Zhou Lin1,Luo Youxi1

Affiliation:

1. School of Science, Hubei University of Technology, Wuhan 430068, China

Abstract

In this study, we propose a filter pruning method based on MCP (Minimax Concave Penalty) regression. The convolutional process is conceptualized as a linear regression procedure, and the regression coefficients serve as indicators to assess the redundancy of channels. In the realm of feature selection, the efficacy of sparse penalized regression gradually outperforms that of Lasso regression. Building upon this insight, MCP regression is introduced to screen convolutional channels, coupled with the coordinate descent method, to effectuate model compression. In single-layer pruning and global pruning analyses, the Top1 loss value associated with the MCP regression compression method is consistently smaller than that of the Lasso regression compression method across diverse models. Specifically, when the global pruning ratio is set to 0.3, the Top1 accuracy of the MCP regression compression method, in comparison with that of the Lasso regression compression method, exhibits improvements of 0.21% and 1.67% under the VGG19_Simple and VGG19 models, respectively. Similarly, for ResNet34, at two distinct pruning ratios, the Top1 accuracy demonstrates enhancements of 0.33% and 0.26%. Lastly, we compare and discuss the novel methods introduced in this study, considering both time and space resource consumption.

Funder

National Social Science Fund of China

National Natural Science Foundation of China

Key Humanities and Social Science Fund of Hubei Provincial Department of Education

Publisher

MDPI AG

Reference43 articles.

1. Li, Y.L. (2022). Model Compression of Deep Neural Networks. [Master’s Thesis, University of Electronic Science and Technology of China].

2. Xu, J.H. (2020). Research on Model Compression and Acceleration of Deep Neural Networks Based on Model Pruning. [Master’s Thesis, Southeast University].

3. Gradient-based learning applied to document recognition;LeCun;Proc. IEEE,1998

4. Research on hyperparameter tuning strategies based on VGG16 network;Zhang;Sci. Innov.,2021

5. Chen, W.J. (2021). Design and Implementation of a High-Speed and High-Precision Matrix Inverter. [Master’s Thesis, Hefei University of Technology].

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3