MAS-Net: Multi-Attention Hybrid Network for Superpixel Segmentation
Author:
Yan Guanghui12, Wei Chenzhen12, Jia Xiaohong13, Li Yonghui12, Chang Wenwen12
Affiliation:
1. School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou 730070, China 2. Key Laboratory of Media Convergence Technology and Communication, Lanzhou 730070, China 3. Key Laboratory of Big Data and Artificial Intelligence in Transportation, Ministry of Education, Beijing Jiaotong University, Beijing 100044, China
Abstract
Superpixels, as essential mid-level image representations, have been widely used in computer vision due to their computational efficiency and redundant compression. Compared with traditional superpixel methods, superpixel algorithms based on deep learning frameworks demonstrate significant advantages in segmentation accuracy. However, existing deep learning-based superpixel algorithms suffer from a loss of details due to convolution and upsampling operations in their encoder–decoder structure, which weakens their semantic detection capabilities. To overcome these limitations, we propose a novel superpixel segmentation network based on a multi-attention hybrid network (MAS-Net). MAS-Net is still based on an efficient symmetric encoder–decoder architecture. First, utilizing residual structure based on a parameter-free attention module at the feature encoding stage enhanced the capture of fine-grained features. Second, adoption of a global semantic fusion self-attention module was used at the feature selection stage to reconstruct the feature map. Finally, fusing the channel with the spatial attention mechanism at the feature-decoding stage was undertaken to obtain superpixel segmentation results with enhanced boundary adherence. Experimental results on real-world image datasets demonstrated that the proposed method achieved competitive results in terms of visual quality and metrics, such as ASA and BR-BP, compared with the state-of-the-art approaches.
Funder
National Natural Science Foundation of China Gansu Provincial Science and Technology Plan Project Key Laboratory of Big Data and Artificial Intelligence in Transportation (Beijing Jiaotong University), Ministry of Education
Reference36 articles.
1. Ren, X., and Malik, J. (2003, January 13–16). Learning a classification model for segmentation. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Nice, France. 2. Kim, S., Park, D., and Shim, B. (2023, January 7–14). Semantic-aware superpixel for weakly supervised semantic segmentation. Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Washington, DC, USA. 3. Superpixel-based fast fuzzy C-means clustering for color image segmentation;Lei;IEEE Trans. Fuzzy Syst.,2019 4. Zhang, S., Ma, Z., Zhang, G., Lei, T., Zhang, R., and Cui, Y. (2020). Semantic image segmentation with deep convolutional neural networks and quick shift. Symmetry, 12. 5. Liu, M., Chen, S., Lu, F., Xing, M., and Wei, J. (2021). Realizing target detection in SAR images based on multiscale superpixel fusion. Sensors, 21.
|
|