Deep Learning Technology to Recognize American Sign Language Alphabet

Author:

Alsharif Bader12,Altaher Ali Salem1ORCID,Altaher Ahmed13ORCID,Ilyas Mohammad1,Alalwany Easa14ORCID

Affiliation:

1. Department of Electrical Engineering and Computer Science, Florida Atlantic University, 777 Glades Road, Boca Raton, FL 33431, USA

2. Department of Computer Science and Engineering, College of Telecommunication and Information, Technical and Vocational Training Corporation (TVTC), Riyadh 11564, Saudi Arabia

3. Electronic Computer Center, Al-Nahrain University, Jadriya, Baghdad 64074, Iraq

4. College of Computer Science and Engineering, Taibah University, Yanbu 46421, Saudi Arabia

Abstract

Historically, individuals with hearing impairments have faced neglect, lacking the necessary tools to facilitate effective communication. However, advancements in modern technology have paved the way for the development of various tools and software aimed at improving the quality of life for hearing-disabled individuals. This research paper presents a comprehensive study employing five distinct deep learning models to recognize hand gestures for the American Sign Language (ASL) alphabet. The primary objective of this study was to leverage contemporary technology to bridge the communication gap between hearing-impaired individuals and individuals with no hearing impairment. The models utilized in this research include AlexNet, ConvNeXt, EfficientNet, ResNet-50, and VisionTransformer were trained and tested using an extensive dataset comprising over 87,000 images of the ASL alphabet hand gestures. Numerous experiments were conducted, involving modifications to the architectural design parameters of the models to obtain maximum recognition accuracy. The experimental results of our study revealed that ResNet-50 achieved an exceptional accuracy rate of 99.98%, the highest among all models. EfficientNet attained an accuracy rate of 99.95%, ConvNeXt achieved 99.51% accuracy, AlexNet attained 99.50% accuracy, while VisionTransformer yielded the lowest accuracy of 88.59%.

Publisher

MDPI AG

Subject

Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry

Reference52 articles.

1. Zwitserlood, I., Verlinden, M., Ros, J., Van Der Schoot, S., and Netherlands, T. (July, January 29). Synthetic signing for the deaf: Esign. Proceedings of the Conference and Workshop on Assistive Technologies for Vision and Hearing Impairment, CVHI, Granada, Spain.

2. World Health Organisation (2022, November 08). Deafness and Hearing Loss. Available online: http://https://www.who.int/news-room/fact-sheets/detail/deafness-and-hearing-loss.

3. Alsaadi, Z., Alshamani, E., Alrehaili, M., Alrashdi, A.A.D., Albelwi, S., and Elfaki, A.O. (2022). A real time Arabic sign language alphabets (ArSLA) recognition model using deep learning architecture. Computers, 11.

4. Alsharif, B., and Ilyas, M. (2022, January 12–13). Internet of Things Technologies in Healthcare for People with Hearing Impairments. Proceedings of the IoT and Big Data Technologies for Health Care: Third EAI International Conference, IoTCare 2022, Virtual Event. Proceedings.

5. Advances in machine translation for sign language: Approaches, limitations, and challenges;Farooq;Neural Comput. Appl.,2021

Cited by 8 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3