Open Access Open Access  Restricted Access Subscription Access

Gesture to Meaning: A Deep Dive into Video Sign Language Recognition

G. Brahmani, R. Sanjana, K. Pranathi, M. Nikesh, M. Bharathi

Abstract


 Automated systems that can recognize and understand sign languages are essential for fostering accessibility and inclusion for the Deaf and Hard of Hearing community. Unfortunately, this community still faces considerable barriers when it comes to communication, education, employment, and full participation in society. Traditional methods for sign language recognition—like rule-based systems or basic machine learning models—often fall short, as they struggle to capture the nuanced and dynamic essence of sign language. Enter Advanced Convolutional Neural Networks (CNNs), which offer a promising way forward by effectively modeling long-range dependencies and temporal sequences in communication. In recent years, innovative deep learning architectures—such as CNNs, Recurrent Neural Networks (RNNs), Multi-Task CNNs (MTCNNs), and transformers—have started to reshape the landscape of sign language recognition. This paper aims to delve into these approaches, comparing them across various metrics like accuracy, dataset size, architecture, and training time. By identifying gaps and potential improvements in video sign language recognition, we hope to enhance these systems and, ultimately, empower the Deaf and Hard of Hearing community to communicate more freely and effectively.


Full Text:

PDF

References


Tunga, Anirudh, Sai Vidyaranya Nuthalapati, and Juan Wachs. “Pose-based sign language recognition using GCN and BERT.” Proceedings of the IEEE/CVF winter conference on applications of computer vision. 2021.

Gunawan, Herman, Narada Thiracitta, and Ariadi Nugroho. “Sign language recognition using modified convolutional neural network model.” 2018 Indonesian Association for Pattern Recognition International Conference (INAPR). IEEE, 2018.

Starner, Thad, and Alex Pentland. “Real-time american sign language recognition from video using hidden markov models.” Proceedings of International Symposium on Computer Vision-ISCV. IEEE, 1995.

Huang, Jie, et al. “Video-based sign language recognition without temporal segmentation.” Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 32. No. 1. 2018.

Koller, Oscar, Jens Forster, and Hermann Ney. “Continuous sign language recognition: Towards large vocabulary statistical recognition systems handling multiple signers.” Computer Vision and Image Understanding 141 (2015): 108-125.

Du, Yao, et al. “Full transformer network with masking future for word-level sign language recognition.” Neurocomputing 500 (2022): 115-123.

Al-Hammadi, Muneer, et al. “Hand gesture recognition for sign language using 3DCNN.” IEEE Access 8 (2020): 79491-79509.

Sridhar, Advaith, et al. “Include: A large scale dataset for indian sign language recognition.” Proceedings of the 28th ACM international conference on multimedia. 2020.

Mindlin, Iván, et al. “A Comparison of Neural Networks for Sign Language Recognition with LSA64.” Conference on Cloud Computing, Big Data Emerging Topics. Cham: Springer International Publishing, 2021.

Xiao, Qinkun, et al. “Multimodal fusion based on LSTM and a couple conditional hidden Markov model for Chinese sign language recognition.” IEEE Access 7 (2019): 112258-112268.


Refbacks

  • There are currently no refbacks.