Preprint / Version 1

An AI-Based Approach To American Sign Language Alphabet Recognition

##article.authors##

  • Vaibhav Akella Polygence

DOI:

https://doi.org/10.58445/rars.3202

Keywords:

Artificial Intelligence (AI), Image Recognition, Computer Vision, American Sign Language, Alphabet Recognition

Abstract

This research presents the development of an AI-powered image recognition model designed to translate American Sign Language (ASL) alphabet gestures into corresponding text characters (A-Z). The models utilize machine learning algorithms, specifically random forest and multi-layer perceptron (MLP) Classifiers with logistic regression as a baseline, trained on a dataset of labelled hand gesture images to classify static signs representing individual letters accurately. Both models achieve high accuracy in identifying hand gestures under controlled lighting and background conditions in a short time frame, with the MLP achieving an accuracy of 98.79% and the random forest achieving a slightly higher accuracy of 99.38% while the baseline achieved an accuracy of 97.66%. These models are particularly beneficial for individuals who are hearing impaired or for those who wish to improve communication with ASL users. Future improvements may include expanding to dynamic gestures as well as recognizing full words or phrases.

References

Akash Nagaraj. (2018). ASL Alphabet [Data set]. Kaggle. https://doi.org/10.34740/KAGGLE/DSV/29550

Bishop, C. M. (2007). Pattern recognition and machine learning. Journal of Electronic Imaging, 16(4), 049901. https://doi.org/10.1117/1.2819119

Breiman, L. (2001). Random Forests. Machine Learning, 45(1), 5–32. https://doi.org/10.1023/a:1010933404324

Mohammadi, M., Chandarana, P., Seekings, J., Hendrix, S., & Zand, R. (2022). Static hand gesture recognition for American sign language using neuromorphic hardware. Neuromorphic Computing and Engineering, 2(4), 044005. https://doi.org/10.1088/2634-4386/ac94f3

Pigou, L., Dieleman, S., Kindermans, P., & Schrauwen, B. (2015). Sign language recognition using convolutional neural networks. In Lecture notes in computer science (pp. 572–578). https://doi.org/10.1007/978-3-319-16178-5_40

Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., & Sutskever, I. (2021). Learning transferable visual models from natural language supervision. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2103.00020

Rokach, L., & Maimon, O. (2006). Decision Trees. In Springer eBooks (pp. 165–192). https://doi.org/10.1007/0-387-25465-x_9

Temitope, A., Nguyen Thanh, H., & Victor, A., 2025. Domain Knowledge in Feature Engineering: Why Human Intuition Still Matters. [online] ResearchGate. https://www.researchgate.net/publication/390492801_Domain_Knowledge_in_Feature_Engineering_Why_Human_Intuition_Still_Matters

Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y., Plu, J., Xu, C., Scao, T.L., Gugger, S., Drame, M., Lhoest, Q. and Rush, A.M., 2020. Transformers: State-of-the-art Natural Language Processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp.38-45. https://doi.org/10.18653/v1/2020.emnlp-demos.6

Downloads

Posted

2025-10-11