Research on Surgical Gesture Recognition in Open Surgery Based on Fusion of R3D and Multi-Head Attention Mechanism
-
Published:2024-09-07
Issue:17
Volume:14
Page:8021
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Men Yutao12, Luo Jian123ORCID, Zhao Zixian123, Wu Hang3, Zhang Guang3, Luo Feng3, Yu Ming3
Affiliation:
1. Tianjin Key Laboratory for Advanced Mechatronic System Design and Intelligent Control, School of Mechanical Engineering, Tianjin University of Technology, Tianjin 300384, China 2. National Demonstration Center for Experimental Mechanical and Electrical Engineering Education, Tianjin University of Technology, Tianjin 300384, China 3. Systems Engineering Institute, Academy of Military Sciences, People’s Liberation Army, Tianjin 300161, China
Abstract
Surgical gesture recognition is an important research direction in the field of computer-assisted intervention. Currently, research on surgical gesture recognition primarily focuses on robotic surgery, with a lack of studies in traditional surgery, particularly open surgery. Therefore, this study established a dataset simulating open surgery for research on surgical gesture recognition in the field of open surgery. With the assistance of professional surgeons, we defined a vocabulary of 10 surgical gestures based on suturing tasks in open procedures. In addition, this paper proposes a surgical gesture recognition method that integrates the R3D network with a multi-head attention mechanism (R3D-MHA). This method uses the R3D network to extract spatiotemporal features and combines it with the multi-head attention mechanism for relational learning of these features. The effectiveness of the R3D-MHA method in the field of open surgery gesture recognition was validated through two experiments: offline recognition and online recognition. The accuracy at the gesture instance level for offline recognition was 92.3%, and the frame accuracy for online recognition was 73.4%. Finally, its performance was further validated on the publicly available JIGSAWS dataset. Compared to other online recognition methods, the accuracy improved without using additional data. This work lays the foundation for research on surgical gesture recognition in open surgery and has significant applications in process monitoring, surgeon skill assessment and educational training for open surgeries.
Reference25 articles.
1. Gao, Y., Vedula, S.S., Reiley, C.E., Ahmidi, N., Varadarajan, B., Lin, H.C., Tao, L., Zappella, L., Béjar, B., and Yuh, D.D. (2014, January 25). JHU-ISI gesture and skill assessment working set (jigsaws): A surgical activity dataset for human motion modeling. Proceedings of the Modeling and Monitoring of Computer Assisted Interventions (M2CAI)—MICCAI Workshop, Boston, MA, USA. 2. Yasar, M.S., and Alemzadeh, H. (July, January 29). Real-time context-aware detection of unsafe events in robot-assisted surgery. Proceedings of the 2020 50th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN), IEEE, Valencia, Spain. 3. Surgical data science for next-generation interventions;Vedula;Nat. Biomed. Eng.,2017 4. Open surgery in the era of minimally invasive surgery;Zhao;Chin. J. Cancer Res.,2022 5. Funke, I., Bodenstedt, S., Oehme, F., von Bechtolsheim, F., and Weitz, J. (2019). Using 3D convolutional neural networks to learn spatiotemporal features for automatic surgical gesture recognition in video. Medical Image Computing and Computer Assisted Intervention—MICCAI 2019, Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019, Springer International Publishing.
|
|