Voxvisor

  • Unique Paper ID: 173613
  • Volume: 11
  • Issue: 10
  • PageNo: 862-870
  • Abstract:
  • The communication barrier between hearing and non-hearing individuals, particularly those with speech impairments, remains a significant challenge. Traditional sign language interpretation methods, often reliant on manual techniques, are time-consuming and limited in their ability to adapt to diverse signing styles. To address these limitations, we propose a deep learning-based system, Voxvisor, for real-time sign language recognition and translation into audible speech. Voxvisor incorporates advanced computer vision techniques, including key-point detection, optical flow, and YOLO (You Only Look Once) feature extraction, to accurately identify and classify sign language gestures. By leveraging deep learning architectures such as CNNs, RNNs, and LSTMs, Voxvisor can effectively learn from a comprehensive dataset of sign language videos, capturing both spatial and temporal characteristics of gestures. Compared to existing manual methods, our approach offers several advantages: real-time recognition, adaptability to various signing styles, improved accuracy. By bridging the communication gap between hearing and non-hearing individuals, Voxvisor has the potential to significantly improve the quality of life for those with speech impairments and promote social inclusion.

Cite This Article

  • ISSN: 2349-6002
  • Volume: 11
  • Issue: 10
  • PageNo: 862-870

Voxvisor

Related Articles