Affiliation:
1. Graduate School of Science and Engineering, Ritsumeikan University, 1-1-1 Noji-higashi, Kusatsu 525-8577, Japan
2. College of Science and Engineering, Ritsumeikan University, 1-1-1 Noji-higashi, Kusatsu 525-8577, Japan
Abstract
Currently, with the rapid development of deep learning, deep neural networks (DNNs) have been widely applied in various computer vision tasks. However, in the pursuit of performance, advanced DNN models have become more complex, which has led to a large memory footprint and high computation demands. As a result, the models are difficult to apply in real time. To address these issues, model compression has become a focus of research. Furthermore, model compression techniques play an important role in deploying models on edge devices. This study analyzed various model compression methods to assist researchers in reducing device storage space, speeding up model inference, reducing model complexity and training costs, and improving model deployment. Hence, this paper summarized the state-of-the-art techniques for model compression, including model pruning, parameter quantization, low-rank decomposition, knowledge distillation, and lightweight model design. In addition, this paper discusses research challenges and directions for future work.
Subject
Computer Networks and Communications,Human-Computer Interaction
Reference151 articles.
1. High Performance CNN Accelerators Based on Hardware and Algorithm Co-Optimization;Yuan;IEEE Trans. Circuits Syst. I Regul. Pap.,2021
2. Barinov, R., Gai, V., Kuznetsov, G., and Golubenko, V. (2023). Automatic Evaluation of Neural Network Training Results. Computers, 12.
3. A review of convolutional neural network architectures and their optimizations;Cong;Artif. Intell. Rev.,2022
4. SAR Target Image Classification Based on Transfer Learning and Model Compression;Zhong;IEEE Geosci. Remote Sens. Lett.,2019
5. Chandio, A., Gui, G., Kumar, T., Ullah, I., Ranjbarzadeh, R., Roy, A.M., Hussain, A., and Shen, Y. (2022). Precise single-stage detector. arXiv.
Cited by
52 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献