Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Sign Language Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras frameworks.

License

NotificationsYou must be signed in to change notification settings

dheerajsharma0401/Automated-Sign-To-Speech-Conversion

Repository files navigation

Automated Sign Language to Speech Conversion system built with OpenCV, Keras/TensorFlow using Deep Learning and Computer Vision concepts in order to communicate using American Sign Language(ASL) based gestures in real-time video streams with differently abled.

                              Pythoncontributions welcomeForksStargazersIssuesLinkedIn

                                   Live Demo

😇 Motivation

A language translator is extensively utilized by the mute people for converting and giving shape to their thoughts. A system is in urgent need ofrecognizing and translating sign language.

Lack of efficient gesture detection system designed specifically for the differently abled, motivates us as a team to do something great in this field. The proposed work aims at converting such sign gestures into speech that can be understood by normal people. The entire model pipeline is developed by CNN architecture for the classification of26 alphabets and one extra alphabet for null character. The proposed work has achieved an efficiency of99.88% .

⚠️ TechStack/framework used

📁 Data Distribution

The dataset used can be downloaded here -Click to Download

This dataset consists of 17113 images belonging to 27 classes:

  • Training Set: 12845 images


Train Data Statistics

  • Test Set: 4368 images


Test Data Statistics

⭐ Features

Our model is capable of predicting gestures from American sign language in real-time with high efficiency. Thesepredicted alphabets are converted to formwords and hence formssentences. These sentences are converted intovoice modules by incorporatingGoogle Text to Speech(gTTS API).

The model is efficient, since we used a compactCNN-based architecture, it’s also computationally efficient and thus making it easier to deploy the model to embedded systems (Raspberry Pi, Google Coral, etc.). This system can therefore be used in real-time applications which aims at bridging the the gap in the process of communication between theDeaf and Dumb people with rest of the world.

🎨 Feature Extraction

  • Gaussian filter is used as a pre-processing technique to make the image smooth and eliminate all the irrelevat noise.
  • Intensity is analyzed and Non-Maximum suppression is implemented to remove false edges.
  • For a better pre-processed image data, double thresholding is implemented to consider only the strong edges in the images.
  • All the weak edges are finally removed and only the strong edges are considered for the further phases.


    The above figure shows pre-processed image with extracted features which is sent to the model for classification.

🎯 Proposed Flow


The figure above shows a detailed pipeline of the model architecture. It can be interpreted that a Convolutional architecture has been proposed.

🔑 Prerequisites

All the dependencies and required libraries are included in the filerequirements.txtSee here

🚀  Installation

  1. Start and fork the repository.

  2. Clone the repo

$ git clone https://github.com/dheerajsharma0401/Automated-Sign-To-Speech-Conversion.git
  1. Change your directory to the cloned repo and create a Python virtual environment named 'test'
$ mkvirtualenv test
  1. Now, run the following command in your Terminal/Command Prompt to install the libraries required
$ pip3 install -r requirements.txt

💡 Working

  1. Open terminal. Go into the cloned project directory and type the following command:
$ python3 jupyter
  1. To train the model, open theASL_train file in jupyter notebook and run all the cells

  2. To detect ASL Gestures in real-time video streams run theASL_Real-Time.ipynb file.

🔑 Results

Our model gave 99.8% accuracy for Sign Language Detection after training viatensorflow-gpu==2.0.0




  • The model has been trained on a python based environment on Jupyter platform.
  • The model is iterated for a total epoch of 20.
  • The model has attained an accuracy of99.88 % accuracy on the Validation set.
  • The prescribed model has been evaluated onTest set where it has attained an accuracy of99.85% with loss of 0.60 %.

We got the following accuracy/loss training curve plot


The above figure shows the Loss plot of the model throughout it's training journey.




The above figure shows the Accuracy plot of the model throughout it's training journey.

📈 Analysis

The training phase data has been evaluated on the Heatmap Plotting representation.


The prescribed analysis aims that each class is most correlated with that traned class.

👏 And it's done!

Feel free to mail me for any doubts/query:email:dheeraj.sharma18@vit.edu

🤝 Contribution

Feel free tofile a new issue with a respective title and description on the theSign-Language-Detection repository. If you already found a solution to your problem,I would love to review your pull request!

❤️ Owner

Made with ❤️  byDheeraj Sharma

👍 Credits

  • https://www.pyimagesearch.com/
  • https://opencv.org/
  • Efthimiou, Eleni & Fotinea, Stavroula-Evita & Vogler, Christian & Hanke, Thomas & Glauert, John & Bowden, Richard & Braffort, Annelies & Collet, Christophe & Maragos, Petros & Segouat, Jérémie. (2009).
  • Sign Language Recognition, Generation, and Modelling: A Research Effort with Applications in Deaf Communication. 21-30. 10.1007/978-3-642-02707-9_3.
  • Pramada, Sawant & Vaidya, Archana. (2013). Intelligent Sign Language Recognition Using Image Processing. IOSR Journal of Engineering. 03. 45-51. 10.9790/3021-03224551.

🤝 Our Contributors

CONTRIBUTORS.md

👀 Code of Conduct

You can find our Code of Conducthere.

👀 License

MIT ©Dheeraj Sharma

About

Sign Language Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras frameworks.

Topics

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors2

  •  
  •  

[8]ページ先頭

©2009-2025 Movatter.jp