This project bridges communication gaps by converting sign language gestures into text and audio, improving accessibility for individuals who use sign language. The system uses real-time gesture recognition and text-to-speech technology to facilitate seamless communication.
- Gesture Recognition: Detects and classifies hand gestures using a CNN model.
- Text-to-Speech Conversion: Converts recognized text into audio.
- Real-time Processing: Processes input from a webcam and generates output in real-time.
- User-friendly Interface: Intuitive GUI for displaying detected characters, suggestions, and sentences.
- Suggestions: Displays word suggestions for improved text accuracy.
- Audio Output: Allows the generated text to be spoken aloud.
First, clone this repository to your local system:
git clone https://github.com/DeltaK17/Sign-Language-to-Text-Converter-with-Audio.git
cd Sign-Language-to-Text-Converter-with-AudioDownload and install Anaconda from the official website.
You can either create a new environment manually or use the provided Conda environment file.
- Create the environment with Python 3.8.20:
conda create -n signlang_env python=3.8.20
- Activate the environment:
conda activate signlang_env
- Install dependencies:
conda install numpy opencv keras matplotlib mediapipe scipy pip install pyttsx3 cvzone pyenchant
If any libraries are missing, install them manually:
conda install numpy opencv keras matplotlib mediapipe scipy
pip install pyttsx3 cvzone pyenchantIf you're using PyCharm for development, follow these steps:
- Open Your Project: Launch PyCharm and open the folder containing your project files.
- Configure the Conda Environment:
- Go to File > Settings (or PyCharm > Preferences on macOS).
- Navigate to Project: <project_name> > Python Interpreter.
- Click the gear icon ⚙️ and select Add....
- Choose Conda Environment and select the
signlang_envyou created.
- Install Missing Packages: Use the PyCharm terminal to install any missing dependencies:
conda install numpy opencv keras matplotlib mediapipe scipy pip install pyttsx3 cvzone pyenchant
- Configure the Run Script:
- Go to Run > Edit Configurations....
- Click + Add New Configuration and select Python.
- Set the script path to
final_pred.py. - Set the working directory to your project folder.
- Select your Conda environment as the interpreter.
- Click the Run button (green triangle) in PyCharm.
- The GUI will launch, allowing you to interact with the application.
- Navigate to your project folder.
- Activate the Conda environment:
conda activate signlang_env
- Run the application:
python final_pred.py
- Python Version: 3.8.20 or python version > 3.8
- Dependencies:
- Mediapipe, NumPy, OpenCV, Keras, Matplotlib, Scipy
- Additional packages: Pyttsx3, Cvzone, PyEnchant
- Hand Tracking: Detects hand landmarks using Mediapipe.
- Gesture Recognition: Uses a CNN model (
cnn8grps_rad1_model.h5) for gesture classification. - GUI: Built with Tkinter for displaying text, suggestions, and output.
- Text-to-Speech: Converts recognized text into audio using Pyttsx3.
├── final_pred.py # Main application script
├── cnn8grps_rad1_model.h5 # Pre-trained model for gesture recognition
├── white.jpg # Background image for processing gestures
├── requirements.txt # pip requirements file
- Support for additional languages and gestures.
- Enhance gesture recognition accuracy.
- Implement dynamic gesture recognition for phrases.
- Optimize processing speed for real-time applications.
This project is licensed under the MIT License. See the LICENSE file for details.