American Sign Language Translator
Problem Statement
- The people having speaking difficulties or hearing disability have a hard time in expressing thoughts and communicating with other people .
- As a result the social, emotional and linguistic growth of such people is hampered and they develop anxiety and depression symptoms.
Solution
- Integrating hand gesture recognition system using computer vision for establishing 2-way communication system.
- To provide information access and services to deaf and dumb people in American sign language.
- The project is scalable and has been extended to capture whole vocabulary of ASL through manual and non manual signs.
Methodology
- Hand Detection using YOLO, CNN and OpenCV.
- FFmpeg to extract hand detected from the video and then Deep CNN was used for mapping the hand gestures recognized to the nearest alphabet in American sign language(ASL).
- Converted the predicted text to speech to make it more interpretable.
IMPLEMENTATION
(ALL ALPHABETS)
1.00
IMPLEMENTATION
(TEST WORD)
Web Implementation
(Using NodeJS for backend and SocketIO for real-time communication)
1.00
Future Goals
-
Real time translation from English to other languages as well.
- Use Facial Detection to extract the user's expression from the video and use it to change tone of predicted speech.
ASL
By Tanishq Saluja
ASL
HackVSIT Presentation
- 104