Abstract
PurposeAccording to the Indian Sign Language Research and Training Centre (ISLRTC), India has approximately 300 certified human interpreters to help people with hearing loss. This paper aims to address the issue of Indian Sign Language (ISL) sentence recognition and translation into semantically equivalent English text in a signer-independent mode.Design/methodology/approachThis study presents an approach that translates ISL sentences into English text using the MobileNetV2 model and Neural Machine Translation (NMT). The authors have created an ISL corpus from the Brown corpus using ISL grammar rules to perform machine translation. The authors’ approach converts ISL videos of the newly created dataset into ISL gloss sequences using the MobileNetV2 model and the recognized ISL gloss sequence is then fed to a machine translation module that generates an English sentence for each ISL sentence.FindingsAs per the experimental results, pretrained MobileNetV2 model was proven the best-suited model for the recognition of ISL sentences and NMT provided better results than Statistical Machine Translation (SMT) to convert ISL text into English text. The automatic and human evaluation of the proposed approach yielded accuracies of 83.3 and 86.1%, respectively.Research limitations/implicationsIt can be seen that the neural machine translation systems produced translations with repetitions of other translated words, strange translations when the total number of words per sentence is increased and one or more unexpected terms that had no relation to the source text on occasion. The most common type of error is the mistranslation of places, numbers and dates. Although this has little effect on the overall structure of the translated sentence, it indicates that the embedding learned for these few words could be improved.Originality/valueSign language recognition and translation is a crucial step toward improving communication between the deaf and the rest of society. Because of the shortage of human interpreters, an alternative approach is desired to help people achieve smooth communication with the Deaf. To motivate research in this field, the authors generated an ISL corpus of 13,720 sentences and a video dataset of 47,880 ISL videos. As there is no public dataset available for ISl videos incorporating signs released by ISLRTC, the authors created a new video dataset and ISL corpus.
Reference72 articles.
1. Vision based hand gesture recognition using dynamic time warping for Indian Sign Language,2016
2. Al-Nafjan, A., Al-Abdullatef, L., Al-Ghamdi, M., Al-Khalaf, N. and Al-Zahrani, W. (2021), “Co-Design of gesture-based Arabic Sign Language (ArSL) recognition”, in Intelligent Human Systems Integration 2021. IHSI 2021. Advances in Intelligent Systems and Computing, Springer, Cham, Vol. 1322, pp. 715-720.
3. Video-based signer independent Arabic Sign Language recognition using hidden Markov models;Applied Soft Computing,2009
4. Towards sign language recognition using EEG-based motor imagery brain computer interface,2017
5. User-independent American Sign Language alphabet recognition based on depth image and PCANet features;IEEE Access,2019
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献