Copyright © 2026 Authors retain the copyright of this article. This article is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
@article{172223,
author = {Debmallya Panja and Aniruddha Das and Irfan Wahid and Arkadyuti Ganguly and Aditya Gupta and Shreyan Dey},
title = {Sign Language Recognizer And Hand Gesture Prediction Using Cnn},
journal = {International Journal of Innovative Research in Technology},
year = {2025},
volume = {11},
number = {8},
pages = {2429-2434},
issn = {2349-6002},
url = {https://ijirt.org/article?manuscript=172223},
abstract = {This sign language search engine uses computer vision and machine learning to instantly recognize and interpret hand gestures. The technology typically uses a webcam or video camera to capture gestures and faces, which are then analyzed using deep learning models that learn from large amounts of data. The project aims to bridge the communication gap between deaf and hard-of-hearing people by translating sign language into written or spoken words, and to improve the accessibility and inclusiveness of various digital interactions. The core elements include a front-end interface where users can interact with the system, a back-end that handles data visualization, and machine learning models that analyze and interpret. The front-end usually has a simple and user-friendly interface, while the back-end handles data flow, processing, and integration with other services (like text-to-speech arguments). Machine learning models are typically based on convolution neural networks (CNNs) or similar models, and are trained on thousands of labeled images or videos to accurately recognize reality. The project can also be designed to provide additional features that can be adapted to different regional languages, such as sign language teaching, language verification, and translation.},
keywords = {Sign Language Recognition, Indian Sign Language (ISL), Image Processing Convolution Neural Networks (CNNs), Human-Computer Interaction (HCI), Static Gesture Recognition, Real-Time Gesture Prediction, Deep Learning, Dataset Augmentation Model Optimization, Real-Time Deployment, Lightweight Models, Mobile Net V2.},
month = {January},
}
Submit your research paper and those of your network (friends, colleagues, or peers) through your IPN account, and receive 800 INR for each paper that gets published.
Join NowNational Conference on Sustainable Engineering and Management - 2024 Last Date: 15th March 2024
Submit inquiry