Author:
Wu Shixiao,Guo Chengcheng,Litifu Ayixiamu,Wang Zhiwei
Abstract
Abstract
Background
For prostate electrosurgery, where real-time surveillance screens are relied upon for operations, manual identification of the prostate capsule remains the primary method. With the need for rapid and accurate detection becoming increasingly urgent, we set out to develop a deep learning approach for detecting the prostate capsule using endoscopic optical images.
Methods
Our method involves utilizing the Simple, Parameter-Free Attention Module(SimAM) residual attention fusion module to enhance the extraction of texture and detail information, enabling better feature extraction capabilities. This enhanced detail information is then hierarchically transferred from lower to higher levels to aid in the extraction of semantic information. By employing a forward feature-by-feature hierarchical fusion network based on the 3D residual attention mechanism, we have proposed an improved single-shot multibox detector model.
Results
Our proposed model achieves a detection precision of 83.12% and a speed of 0.014 ms on NVIDIA RTX 2060, demonstrating its effectiveness in rapid detection. Furthermore, when compared to various existing methods including Faster Region-based Convolutional Neural Network (Faster R-CNN), Single Shot Multibox Detector (SSD), EfficientDet and others, our method Attention based Feature Fusion Single Shot Multibox Detector (AFFSSD) stands out with the highest mean Average Precision (mAP) and faster speed, ranking only below You Only Look Once version 7 (YOLOv7).
Conclusions
This network excels in extracting regional features from images while retaining the spatial structure, facilitating the rapid detection of medical images.
Publisher
Springer Science and Business Media LLC
Reference31 articles.
1. Xiao J, Wu Y, Chen Y, Wang S, Wang Z, Ma J. LSTFE-Net: long short-term feature enhancement network for video small object detection. In: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada. 2023. pp. 14613–22.
2. Xiao J, Guo H, Zhou J, Zhao T, Yu Q, Chen Y, et al. Tiny object detection with context enhancement and feature purification. Expert Syst Appl. 2023;211:118665. https://doi.org/10.1016/j.eswa.2022.118665.
3. Yu J, Zhang C, Wang H, Zhang D, Song Y, Xiang T et al. 3D medical point transformer: introducing convolution to attention networks for medical point cloud analysis. arXiv preprint arXiv:2112.04863. 2021.
4. Shi T, Jiang H, Zheng B. C2MA-Net: cross-modal cross-attention network for acute ischemic stroke lesion segmentation based on CT perfusion scans. IEEE Trans Biomed Eng. 2022;69(1):108–18. https://doi.org/10.1109/tbme.2021.3087612.
5. Duran A, Dussert G, Rouvière O, Jaouen T, Jodoin P-M, Lartizien C. ProstAttention-Net: a deep attention model for prostate cancer segmentation by aggressiveness in MRI scans. Med Image Anal. 2022;77:102347. https://doi.org/10.1016/j.media.2021.102347.