Affiliation:
1. Department of Electronics and Information Engineering, Bozhou University, Bozhou, AnHui, China
2. School of Computer Science and Engineering, Xi’an University of Technology, Xi’an, Shaanxi, China
3. School of Artificial Intelligence, Hefei University of Technology, Hefei, Anhui, China
Abstract
The performance of a convolutional neural network (CNN) model is influenced by several factors, such as depth, width, network structure, size of the receptive field, and feature map scaling. The optimization of the best combination of these factors poses as the main difficulty in designing a viable architecture. This article presents an analysis of key factors influencing network performance, offers several strategies for constructing an efficient convolutional network, and introduces a novel architecture named TbsNet (thin-branch structure network). In order to minimize computation costs and feature redundancy, lightweight operators such as asymmetric convolution, pointwise convolution, depthwise convolution, and group convolution are implemented to further reduce the network’s weight. Unlike previous studies, the TbsNet architecture design rejects the reparameterization method and adopts a plain, simplified structure which eliminates extraneous branches. We conduct extensive experiments, including network depth, width, etc. TbsNet performs well on benchmark platforms, Top 1 Accuracy on CIFAR-10 is 97.02%, on CIFAR-100 is 83.56%, and on ImageNet-1K is 86.17%. Tbs-UNet’s DSC on the Synapse dataset is 78.39%, higher than TransUNet’s 0.91%. TbsNet can be competent for some downstream tasks in computer vision, such as medical image segmentation, and thus is competitive with prior state-of-the-art deep networks such as ResNet, ResNeXt, RepVgg, ParNet, ConvNeXt, and MobileNet.
Funder
First Batch of Industry-University Cooperation Collaborative Education Projects in 2021
Natural Science Foundation of Colleges and Universities of Anhui Province
Excellent top-of-the-line Talent Training Program of Anhui Province Colleges and Universities
Reference52 articles.
1. DenseNet models for tiny ImageNet classification;Abai;ArXiv,2019
2. TransUNet: transformers make strong encoders for medical image segmentation;Chen,2021
3. A survey of model compression and acceleration for deep neural networks;Cheng;IEEE Signal Processing Magazine,2020
4. Approximation by superpositions of a sigmoidal function;Cybenko;Mathematics of Control, Signals, and Systems,1992
5. A guide to receptive field arithmetic for convolutional neural networks | by Dang Ha The Hien | ML Review | Medium. 1–8;Dang Ha The Hien,2017