A real-time application for detecting and recognizing sign language gestures using a webcam feed.
- Aim
- Project Overview
- Demo
- Features
- Getting Started
- Usage
- Project Report
- Contributing
- License
- Contact
This project aims to create a sign language translator using machine learning techniques and Python programming. The application utilizes various modules, primarily Mediapipe, Landmark, and Random Forest algorithms to interpret and translate sign language gestures into text or spoken language.
Sign language is a crucial form of communication for individuals with hearing impairments. This project focuses on bridging the communication gap by creating a tool that can interpret sign language gestures in real-time and convert them into understandable text or speech.
This project leverages Flask for the web interface and TensorFlow/Keras for the machine learning model to recognize sign language gestures in real-time from a webcam feed.
American Sign Language Convention for Alphabets.
Custom Sign Language for Words / Sentences.
Showcasing a demonstration of the Realtime Sign Language Detection
sign2text_demo.mp4
- Real-time sign language recognition: Captures hand gestures using the Mediapipe library to track landmarks and movements.
- Landmark analysis: Utilizes Landmark module to extract key points and gestures from hand movements.
- Machine learning translation: Employs Random Forest algorithm to classify and interpret gestures into corresponding text.
- Text-to-speech: For better communication the text can be converted to spoken language using the speech synthesis.
To get started with the Sign Language Translator, follow these steps:
- Python: Provides a vast array of libraries and frameworks for machine learning, computer vision, and data processing.
- TensorFlow: For building and training machine learning models.
- Scikit-learn: For implementing the Random Forest algorithm for sign language recognition.
- Numpy: For numerical computations and data manipulation.
- Mediapipe: For real-time hand tracking and landmark detection.
- OpenCV: For video processing and computer vision tasks.
- Flask: Web framework to develop the application.
- Flask-SocketIO: Adds low-latency bi-directional communication between clients and the server to Flask applications.
- Clone the repository:
git clone https://github.com/uzibytes/sign2text.git
cd sign2text
- Create and activate a virtual environment:
python -m venv venv
source venv/bin/activate # On Windows use `venv\Scripts\activate`
- Install required libraries:
pip install -r requirements.txt
- Ensure a webcam is connected to your system.
-
Start the Flask application:
python app.py
-
Open your web browser and navigate to :
http://127.0.0.1:5000/
-
The web interface will display the webcam feed and detected sign language gestures.
For detailed insights, analysis, and findings, refer to the Project Report provided in the repository or click here.
Contributions are welcome! If you'd like to contribute to this project, feel free to open issues, create pull requests, or reach out to discuss potential improvements.
This section states that the Realtime Sign Language Detection Using LSTM Model project is released under the MIT License. It briefly describes the terms and conditions of the license, such as the permission to use, modify, and distribute the project, with appropriate attribution. It provides a link to the full text of the MIT License for further reference.
This is a Final Year B.Tech Project for the session 2020-24. This project is completed under the Guidance of Dr. Shashi Raj (Assistant Professor, Dept. of CSE, Bakhtiyarpur College of Engineering, Patna). This is a group project and the members are :
- Ujjwal Raj - 20105126034
- Krishna Raj - 20105126040
- Prashant Kumar - 20105126043
- Rajnish Puri - 20105126031
For any questions or inquiries, feel free to contact.