It is very difficult for dumb (speechless) people to interact with normal people because most people are not able to understand sign language. We are developing a smart glove that can translate hand gestures into speech. We are integrating flex sensors and Mpu6050 (gyroscope + accelerometer) wi
hand gesture to speech conversion for dumb people
It is very difficult for dumb (speechless) people to interact with normal people because most people are not able to understand sign language. We are developing a smart glove that can translate hand gestures into speech. We are integrating flex sensors and Mpu6050 (gyroscope + accelerometer) with raspberry pi. Data of hand gestures are collected from these sensors and then this data is processed. After preprocessing different machine learning algorithms (SVM, KNN, DTC, ANN) are used to train the model and finally compare all these algorithms to find out which is most suitable, and then the trained model is tested in real-time to view its accuracy and sustainability. It is a multi-purpose glove that finds its applications in controlling digital devices, robotics, Virtual Reality, and many others.
Project Implementation Method:
Working methodology of our project is described as:
1. Gesture recording:
Gestures are captured using flex sensors and MPU 6050 (gyroscope + accelerometer). The flex sensor works as a voltage divider when connected with a resistor. It tells us about the bend of fingers in the form of analog values. So, we have to convert it to digital form so that our micro-controller can understand it. So, an analog to digital converter is used to make this information valuable to a microcontroller. Moreover, MPU-6050 will be used for the orientation of the hand in space. Thus, the values of MPU-6050 will be conveyed to the microcontroller. Thus, the flex sensor will give the values but the gyroscope will determine its orientation in space and gesture will be recorded accordingly.
2. Data pre-processing:
Null values and stray values are removed from the dataset to make the data clean and valuable. Moreover, data is split into training and testing with a ratio of 70:30.
3. Classification of gestures (using ML algorithms):
After getting cleaned data different classifiers are implemented. Decision Tree Classifier gives an accuracy of 95.07%. K Nearest Neighbor gives an accuracy of 96.23%. While Artificial Neural Network gives an accuracy of 98.112%. So, the classifier returning the highest accuracy is implemented.
4. Text to speech conversion:
When a gesture is recognized, its respective speech is its output in the form of text. Now, this text is converted into speech using Python Text to Speech (pyttsx) library. This creates a .wav extension file that can be played by any speaker. So, a speaker is attached with Bluetooth to play voice files of gestures. So, when a gesture is recognized its voice file is automatically played. In this way, a gesture is converted into listenable sound.
Smart glove converts sign language to speech
It can be used in Virtual reality applications.
It can control different electronic devices with gestures.
It can manipulate a robotic arm
| Attributes | Explanation in the context of application |
| Input | The input is obtained from hand Gestures using flex sensors and MPU 6050 (gyroscope + accelerometer). It tells us about bend of fingers in the form of analog values. So, analog to digital converter MCP 3008 is used to make this information readable to micro controller. Moreover MPU-6050 is used for the orientation of hand in space. Thus, flex sensor will give gesture values but gyroscope will determine its orientation in space and gesture will be recorded accordingly. |
| Output | Speech obtained from the speaker |
| No. of gestures recorded. | Twenty Pakistani sign language gesture are recorded and trained. |
| No of samples | There are almost hundred samples are recorded per gestures. So the sampling rate is 100. |
| Accuracy | Decision Tree Classifier gives an accuracy of 95.07%. K Nearest Neighbour gives an accuracy of 96.23%. While Artificial Neural Network gives an accuracy of 98.112%. |
| Components | Flex Sensor, MCP 3008, MPU 6050, Raspberry pi, Zong Internet Device, Speaker, Power Bank, PCB |
Attributes
Input
Output
No. of gestures recorded.
No of samples
Accuracy
Components
| Elapsed time in (days or weeks or month or quarter) since start of the project | Milestone | Deliverable |
|---|---|---|
| Month 1 | Literature Review | succeeded |
| Month 2 | Model and circuit Design | succeeded |
| Month 3 | Model Training | succeeded |
| Month 4 | Model Testing | succeeded |
| Month 5 | Improvement | succeeded |
| Month 6 | Implementing max accuracy classifier | succeeded |
| Month 7 | Demonstration | succeeded |
| Month 8 | Project report | In progress |
The project aims to solve the local Sugar industry problem of ethanol-water separation at...
Simultaneous Localisation and Mapping (SLAM), a technology which allows a device...
A wireless Bioradar sensor is based on a Doppler theory.The radar set differentiate betwee...
In our project we are going to study about block codes and convolutional codes and impleme...
The purpose of this project is to increase the power and efficiency of mono-crystalline PV...