Author:
Zhang Yidan,Wang Guangjin,Yang Taibo,Pang Tianfeng,He Zhenan,Lv Jiancheng
Abstract
AbstractRecently, many studies have been carried out on model compression to handle the high computational cost and high memory footprint brought by the implementation of deep neural networks. In this paper, model compression of convolutional neural networks is constructed as a multiobjective optimization problem with two conflicting objectives, reducing the model size and improving the performance. A novel structured pruning method called Conventional-based and Evolutionary Approaches Guided Multiobjective Pruning (CEA-MOP) is proposed to address this problem, where the power of conventional pruning methods is effectively exploited for the evolutionary process. A delicate balance in pruning rate and model accuracy has been automated achieved by a multiobjective optimization evolutionary model. First, an ensemble framework integrates pruning metrics to establish a codebook for further evolutionary operations. Then, an efficient coding method is developed to shorten the length of chromosome, thus ensuring its superior scalability. Finally, sensitivity analysis is automatically carried out to determine the upper bound of pruning rate for each layer. Notably, on CIFAR-10, CEA-MOP reduces more than 50% FLOPs on ResNet-110 and improves the relative accuracy. Moreover, on ImageNet, CEA-MOP reduces more than 50% FLOPs on ResNet-101 with negligible top-1 accuracy drop.
Funder
the National Natural Science Foundation of China
the Key Research and Development Project of Sichuan
the National Key Research and Development Project of China
the State Key Program of National Science Foundation of China
the National Natural Science Fund for Distinguished Young Scholar
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Software
Reference64 articles.
1. Bau D, Zhou B, Khosla A, Oliva A, Torralba A (2017) Network dissection: quantifying interpretability of deep visual representations. In: CVPR
2. Bèack T (1998) Evolutionary algorithms in theory and practice: evolution strategies, evolutionary programming. Genetic Algorithms/T, Bäck
3. Cheng Y, Wang D, Zhou P, Zhang T (2017) A survey of model compression and acceleration for deep neural networks. arXiv:171009282
4. Chiu W, Yen GG, Juan T (2016) Minimum Manhattan distance approach to multiple criteria decision making in multiobjective optimization problems. IEEE Trans Evol Comput 20(6):972–985
5. Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE Trans Evol Comput 6(2):182–197
Cited by
10 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献