An AI-Based Approach To American Sign Language Alphabet Recognition
DOI:
https://doi.org/10.58445/rars.3202Keywords:
Artificial Intelligence (AI), Image Recognition, Computer Vision, American Sign Language, Alphabet RecognitionAbstract
This research presents the development of an AI-powered image recognition model designed to translate American Sign Language (ASL) alphabet gestures into corresponding text characters (A-Z). The models utilize machine learning algorithms, specifically random forest and multi-layer perceptron (MLP) Classifiers with logistic regression as a baseline, trained on a dataset of labelled hand gesture images to classify static signs representing individual letters accurately. Both models achieve high accuracy in identifying hand gestures under controlled lighting and background conditions in a short time frame, with the MLP achieving an accuracy of 98.79% and the random forest achieving a slightly higher accuracy of 99.38% while the baseline achieved an accuracy of 97.66%. These models are particularly beneficial for individuals who are hearing impaired or for those who wish to improve communication with ASL users. Future improvements may include expanding to dynamic gestures as well as recognizing full words or phrases.
References
Akash Nagaraj. (2018). ASL Alphabet [Data set]. Kaggle. https://doi.org/10.34740/KAGGLE/DSV/29550
Bishop, C. M. (2007). Pattern recognition and machine learning. Journal of Electronic Imaging, 16(4), 049901. https://doi.org/10.1117/1.2819119
Breiman, L. (2001). Random Forests. Machine Learning, 45(1), 5–32. https://doi.org/10.1023/a:1010933404324
Mohammadi, M., Chandarana, P., Seekings, J., Hendrix, S., & Zand, R. (2022). Static hand gesture recognition for American sign language using neuromorphic hardware. Neuromorphic Computing and Engineering, 2(4), 044005. https://doi.org/10.1088/2634-4386/ac94f3
Pigou, L., Dieleman, S., Kindermans, P., & Schrauwen, B. (2015). Sign language recognition using convolutional neural networks. In Lecture notes in computer science (pp. 572–578). https://doi.org/10.1007/978-3-319-16178-5_40
Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. (2021). Learning transferable visual models from natural language supervision. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2103.00020
Rokach, L., & Maimon, O. (2006). Decision Trees. In Springer eBooks (pp. 165–192). https://doi.org/10.1007/0-387-25465-x_9
Temitope, A., Nguyen Thanh, H., & Victor, A., 2025. Domain Knowledge in Feature Engineering: Why Human Intuition Still Matters. [online] ResearchGate. https://www.researchgate.net/publication/390492801_Domain_Knowledge_in_Feature_Engineering_Why_Human_Intuition_Still_Matters
Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T.L., Gugger, S., Drame, M., Lhoest, Q. and Rush, A.M., 2020. Transformers: State-of-the-art Natural Language Processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp.38-45. https://doi.org/10.18653/v1/2020.emnlp-demos.6
Downloads
Posted
Categories
License
Copyright (c) 2025 Vaibhav Akella

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.