- Notifications
You must be signed in to change notification settings - Fork4
Sign Language Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras frameworks.
License
dheerajsharma0401/Automated-Sign-To-Speech-Conversion
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
A language translator is extensively utilized by the mute people for converting and giving shape to their thoughts. A system is in urgent need ofrecognizing and translating sign language.
Lack of efficient gesture detection system designed specifically for the differently abled, motivates us as a team to do something great in this field. The proposed work aims at converting such sign gestures into speech that can be understood by normal people. The entire model pipeline is developed by CNN architecture for the classification of26 alphabets and one extra alphabet for null character. The proposed work has achieved an efficiency of99.88% .
The dataset used can be downloaded here -Click to Download
This dataset consists of 17113 images belonging to 27 classes:
- Training Set: 12845 images
- Test Set: 4368 images
Our model is capable of predicting gestures from American sign language in real-time with high efficiency. Thesepredicted alphabets are converted to formwords and hence formssentences. These sentences are converted intovoice modules by incorporatingGoogle Text to Speech(gTTS API).
The model is efficient, since we used a compactCNN-based architecture, it’s also computationally efficient and thus making it easier to deploy the model to embedded systems (Raspberry Pi, Google Coral, etc.). This system can therefore be used in real-time applications which aims at bridging the the gap in the process of communication between theDeaf and Dumb people with rest of the world.
- Gaussian filter is used as a pre-processing technique to make the image smooth and eliminate all the irrelevat noise.
- Intensity is analyzed and Non-Maximum suppression is implemented to remove false edges.
- For a better pre-processed image data, double thresholding is implemented to consider only the strong edges in the images.
- All the weak edges are finally removed and only the strong edges are considered for the further phases.
The above figure shows pre-processed image with extracted features which is sent to the model for classification.
The figure above shows a detailed pipeline of the model architecture. It can be interpreted that a Convolutional architecture has been proposed.
All the dependencies and required libraries are included in the filerequirements.txt
See here
Start and fork the repository.
Clone the repo
$ git clone https://github.com/dheerajsharma0401/Automated-Sign-To-Speech-Conversion.git
- Change your directory to the cloned repo and create a Python virtual environment named 'test'
$ mkvirtualenv test
- Now, run the following command in your Terminal/Command Prompt to install the libraries required
$ pip3 install -r requirements.txt
- Open terminal. Go into the cloned project directory and type the following command:
$ python3 jupyter
To train the model, open theASL_train file in jupyter notebook and run all the cells
To detect ASL Gestures in real-time video streams run theASL_Real-Time.ipynb file.
- The model has been trained on a python based environment on Jupyter platform.
- The model is iterated for a total epoch of 20.
- The model has attained an accuracy of99.88 % accuracy on the Validation set.
- The prescribed model has been evaluated onTest set where it has attained an accuracy of99.85% with loss of 0.60 %.
Feel free to mail me for any doubts/query:email:dheeraj.sharma18@vit.edu
Feel free tofile a new issue with a respective title and description on the theSign-Language-Detection repository. If you already found a solution to your problem,I would love to review your pull request!
Made with ❤️ byDheeraj Sharma
- https://www.pyimagesearch.com/
- https://opencv.org/
- Efthimiou, Eleni & Fotinea, Stavroula-Evita & Vogler, Christian & Hanke, Thomas & Glauert, John & Bowden, Richard & Braffort, Annelies & Collet, Christophe & Maragos, Petros & Segouat, Jérémie. (2009).
- Sign Language Recognition, Generation, and Modelling: A Research Effort with Applications in Deaf Communication. 21-30. 10.1007/978-3-642-02707-9_3.
- Pramada, Sawant & Vaidya, Archana. (2013). Intelligent Sign Language Recognition Using Image Processing. IOSR Journal of Engineering. 03. 45-51. 10.9790/3021-03224551.
You can find our Code of Conducthere.
MIT ©Dheeraj Sharma
About
Sign Language Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras frameworks.
Topics
Resources
License
Code of conduct
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Contributors2
Uh oh!
There was an error while loading.Please reload this page.