Affiliation:
1. School of Computing, SASTRA Deemed University, Thanjavur, India
2. Novosibirsk State Technical University, Russian Federation
3. Symbiosis Institute of Computer Studies and Research, Symbiosis International (Deemed University), Pune, India
4. Symbiosis Centre for Applied Artificial Intelligence, Symbiosis International (Deemed University), Pune, India
Abstract
The Sign Language Recognition system intends to recognize the Sign language used by the hearing and vocally impaired populace. The interpretation of isolated sign language from static and dynamic gestures is a difficult study field in machine vision. Managing quick hand movement, facial expression, illumination variations, signer variation, and background complexity are amongst the most serious challenges in this arena. While deep learning-based models have been used to accomplish the entirety of the field's state-of-the-art outcomes, the previous issues have not been fully addressed. To overcome these issues, we propose a Hybrid Neural Network Architecture for the recognition of Isolated Indian and Russian Sign Language. In the case of static gesture recognition, the proposed framework deals with the 3D Convolution Net with an atrous convolution mechanism for spatial feature extraction. For dynamic gesture recognition, the proposed framework is an integration of semantic spatial multi-cue feature detection, extraction, and Temporal-Sequential feature extraction. The semantic spatial multi-cue feature detection and extraction module help in the generation of feature maps for Full-frame, pose, face, and hand. For face and hand detection, GradCam and Camshift algorithm have been used. The temporal and sequential module consists of a modified auto-encoder with a GELU activation function for abstract high-level feature extraction and a hybrid attention layer. The hybrid attention layer is an integration of segmentation and spatial attention mechanism. The proposed work also involves creating a novel multi-signer, single, and double-handed Isolated Sign representation dataset for Indian and Russian Sign Language. The experimentation was done on the novel dataset created. The accuracy obtained for Static Isolated Sign Recognition was 99.76%, and the accuracy obtained for Dynamic Isolated Sign Recognition was 99.85%. We have also compared the performance of our proposed work with other baseline models with benchmark datasets, and our proposed work proved to have better performance in terms of Accuracy metrics.
Funder
Department of Science & Technology (DST), India
Publisher
Association for Computing Machinery (ACM)
Reference61 articles.
1. Y. Saleh and G. Issa. 2020. Arabic sign language recognition through deep neural networks fine-tuning. https://www.learntechlib.org/p/217934/.
2. K. Wangchuk K. Wangchuk and P. Riyamongkol. 2020. Bhutanese sign language hand-shaped alphabets and digits detection and recognition (doctoral dissertation naresuan university). http://nuir.lib.nu.ac.th/dspace/handle/123456789/2491.
3. An eight-layer convolutional neural network with stochastic pooling, batch normalization and dropout for fingerspelling recognition of Chinese sign language
4. Turkish sign language digits classification with CNN using different optimizers;Sevli O.;Int. Adv. Res. Eng. J.,2020
5. R. Elakkiya and E. Rajalakshmi Islan. Mendeley Data Vol. 1. https://data.mendeley.com/datasets/rc349j45m5/1.
Cited by
20 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献