Cross-Attention-Guided Feature Alignment Network for Road Crack Detection
-
Published:2023-09-19
Issue:9
Volume:12
Page:382
-
ISSN:2220-9964
-
Container-title:ISPRS International Journal of Geo-Information
-
language:en
-
Short-container-title:IJGI
Author:
Xu Chuan1, Zhang Qi1, Mei Liye1, Chang Xiufeng2, Ye Zhaoyi1ORCID, Wang Junjian3, Ye Lang3, Yang Wei3ORCID
Affiliation:
1. School of Computer Science, Hubei University of Technology, Wuhan 430068, China 2. Unit 92493, Huludao 125000, China 3. School of Information Science and Engineering, Wuchang Shouyi University, Wuhan 430064, China
Abstract
Road crack detection is one of the important issues in the field of traffic safety and urban planning. Currently, road damage varies in type and scale, and often has different sizes and depths, making the detection task more challenging. To address this problem, we propose a Cross-Attention-guided Feature Alignment Network (CAFANet) for extracting and integrating multi-scale features of road damage. Firstly, we use a dual-branch visual encoder model with the same structure but different patch sizes (one large patch and one small patch) to extract multi-level damage features. We utilize a Cross-Layer Interaction (CLI) module to establish interaction between the corresponding layers of the two branches, combining their unique feature extraction capability and contextual understanding. Secondly, we employ a Feature Alignment Block (FAB) to align the features from different levels or branches in terms of semantics and spatial aspects, which significantly improves the CAFANet’s perception of the damage regions, reduces background interference, and achieves more precise detection and segmentation of damage. Finally, we adopt multi-layer convolutional segmentation heads to obtain high-resolution feature maps. To validate the effectiveness of our approach, we conduct experiments on the public CRACK500 dataset and compare it with other mainstream methods. Experimental results demonstrate that CAFANet achieves excellent performance in road crack detection tasks, which exhibits significant improvements in terms of F1 score and accuracy, with an F1 score of 73.22% and an accuracy of 96.78%.
Funder
Hubei University of Technology Natural Science Foundation of Hubei Province University Student innovation and Entrepreneurship Training Program Project
Subject
Earth and Planetary Sciences (miscellaneous),Computers in Earth Sciences,Geography, Planning and Development
Reference54 articles.
1. Vibration vs. vision: Best approach for automated pavement distress detection;Lekshmipathy;Int. J. Pavement Res. Technol.,2020 2. Alfarrarjeh, A., Trivedi, D., Kim, S.H., and Shahabi, C. (2018, January 10–13). A deep learning approach for road damage detection from smartphone images. Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA. 3. Fang, K., Ouyang, J., and Hu, B. (2021). Swin-HSTPS: Research on target detection algorithms for multi-source high-resolution remote sensing images. Sensors, 21. 4. Arya, D., Maeda, H., Ghosh, S.K., Toshniwal, D., Mraz, A., Kashiyama, T., and Sekimoto, Y. (2020). Transfer learning-based road damage detection for multiple countries. arXiv. 5. CNN based road user detection using the 3D radar cube;Palffy;IEEE Robot. Autom. Lett.,2020
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|