Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Rahul Sesha Sai Vemparala, Malarselvi S, Vara Prasad Koppineedi, Bhagyasri Burada, Lokesh Raj Kavali
DOI Link: https://doi.org/10.22214/ijraset.2023.50284
Certificate: View Certificate
There are billions of people in the world who have hearing and speaking impairments. According to a survey roughly 1 billion people are born with hearing and speaking problems. Such disabled or challenged people struggle a lot to communicate with normal humans. A blessing to them is Sign Language, using which they can communicate efficiently with others and can express their ideas and feelings. Sign language is a procedure in which people communicate through making gestures with their hand which are globally recognized and have standard interpretation. Sign Language can be taught to everyone, but it requires a lot of trainers and experts to train everyone personally. Since technology has a lot of reach and resources in this modern era, imbibing such Mechanisms into technology increases the reach of Sign Language Exponentially. Being Computer and Software Engineers our main aim is to use our skills and knowledge of current technologies and create a Sign Language And Gesture interpreter that helps to recognise arm, wrist and finger gestures, emotional expressions and behavioural sign patterns and the observation of movements in different partitions of a person’s body and helps communicate and express one’s thought clearly.Hence we bridging the gap between them and raise awareness. SLR helps in translating sign languages and gestures into other forms of communication like text and speech in order to let deaf people and mute people communicate with one another. While having a significant social influence, this task is very complicated to implement due to high ambiguity and more number of Sign Language Standards being used by various people. Current SLR techniques create classification models based on hand-crafted characteristics to describe sign language motion. Nonetheless, it is challenging to build trustworthy features that can detect and predict various gestures from various types of people. Our model will be able to extract various anchor points present on the body of people along with facial expressions being taken into consideration for a better output. To improve performance, the 3D CNN is fed multiple channels of video feeds that include colour, depth, and trajectory information as well as body joint locations and depth clues. We illustrate the proposed model\'s superiority to the conventional methods based on hand-crafted features by validating it on a real dataset acquired using Microsoft Kinect.
I. INTRODUCTION
There are various types and standards of sign languages present throughout the world for aiding the people with hearing and speaking disability to communicate with the world. Basically we need to capture the movements and hand gestures made by the people as input through Webcam and process the data into frames and these frames are compared to the tested dataset to obtain the result. According to statistics and the information present there are more than 300 sign languages which are practised and used by millions of people. The way in which the particular sign languages are designed is dependent on the culture of the regions, history, mindset of people. The most frequently used language and highly popular sign language is ASL(American Sign Language). Even India has its own Sign Language know as ISL or Indian Sign Language. Like common languages for suppose English or Hindi, each language has its own vocabulary and grammar. Similarly even Sign Language has its own type of structure and grammar. Both these Sign Languages try to convey the message through hand movements and analysis of these movements, for better understanding and more complex sentence structures we have to include facial expressions. ISL was recently proposed and it came into action far later than ASL. That is why still most of the people around the world use American Sign language as a Standard.
The World Health Organisation (WHO) claims that Over a billion of the population who are deaf and can’t speak were able to learn and efficiently communicate using the American Sign Language. Every language be it a spoken on or a Signalling Language has a Syntax and Structure.
Factors that Determine Sign Language Recognition include:
Our Software Built by Group No:4 B12 of Gitam University will help impaired people understand what we are trying to communicate to them using the particular gesture, by showing the output of each gesture we perform in the form of a text. Hence the impaired person once after looking at the gesture or sign language movement will also see the output of the sign language hence understanding the meaning of the particular gesture.
In our Model of the Project we are using the Standard of ASL (American Sign Language) as the input, these are few standard gestures used throughout the world.
In our model which is still under construction and Improvement we are using VGG16 Machine Learning Structure aids us in producing more clear and accurate results of prediction.
VGG16 is one of the most popular Deep learning Architectures used for implementing various Image Classification and Computer Vision Applications and has proved to be successful.
It Dominates other similar softwares like Normalization Architecture and Pooling Architecture which comes out to be more efficient and accurate.
a. CNN (Convolutional Neural Networks)
b. American Sign Language(ASL)
c. Machine Learning
II. PROBLEM STATEMENT AND OBJECTIVE
A. Objective
Mute people find it incredibly challenging to communicate with normal people. because hand sign language is not taught to the general public. It is exceedingly challenging to get their point through during an emergency. The conversion of sign language into human hearing voice is the answer to this issue. The CNN methods for detecting hand motion or gesture.
B. Problem Statement
Due to birth abnormalities, accidents, and oral infections, there has been a sharp rise in the number of people who are deaf and dumb in recent years. Deaf and dumb persons must rely on some form of visual communication since they are unable to communicate with regular people. Across the world, many different languages are spoken and translated. The term "Special People" refers to persons who have difficulties hearing and speaking. "The Dumb" and "The Deaf" people, respectively, have difficulty understanding what the other person is attempting to say. Occasionally individuals will misinterpret these communications using sign language, lip reading, or lip sync.
III. LITERATURE SURVEY
A. Medium of Communication and translation for deaf and mute people with other people. Boon for the Impaired!
The ability to express oneself by reacting to events taking place around him is one of nature's most priceless gifts to humans. Every typical person observes, listens, and then responds to the circumstances by speaking up. Nevertheless, the population which comes under below poverty line or economically weaker sections are being oppressed and are not granted access to such valuable resources which are not meant to be commercial.
This widens the divide between privileged people and average people. They can speak with each other more easily thanks to this software. The input can be in the form of jpeg or video and can also constitute of NLP or natural Language Processing and Conversions like Text to speech included with it. Frame construction from videos, autoscaling, auto focus and image frame size adjustment and image syncing with linguistic or Language knowledge bases will all be part of the Procedure. Relevant audio is then generated using the Google Text-to-Speech(TTS) API. The text is then further mapped to pertinent animated gestures from the database.
B. Collection micro information and implementing it in the larger scale and performing mathematical operations for the output.
Orientation sensing is regarded as a crucial tool for implementing algorithms which are used in check shapes and structures involving geometric calculations and also determine how the agent is interacting with the environment and how it is responding to stimuli and performing actions. Major Indulging Factors involve orientation, spacial coordinates and dynamics etc. The framework definition elaborates a collection of gestures that are not particular to any one application and is independent of categorization techniques and sensor technologies. It supports a configurable number of sensors and allows for interoperability between several sensors. A key element of the framework is all he datasets we collect for the training and testing purposes, these largely determine the accuracy and till what extent the model can predict the inputs. A framework for developing gesture-based interactions is created using multiple stimulated anchor points where the points wont move and the objects connected to these points move which inturn enable us to detect the gesture. The application of this framework is illustrated through the creation of real, hardware- and software-based remote controllers for media players. The whole testing experiment is performed in such a way that the setup is prepared forward for American sign language (ASL) exhibiting the sign is played on the monitor of a portable computer, and the words that go with them are taken from the American sign language dictionary. The word is finger spelled if there is no symbol for it in the vocabulary of signs. In the actual world, deaf people perform this for words without distinct signals, such as proper names.These models also include the most integral part of the machine learning study which are: Supervised, Unsupervised and Reinforcement Learning. The suggested job is an addition to existing study into the "Boltay Haath" signal, which is a finger movement used by people with voice impairments to communicate. The suggested AISR system, when merged with the Boltay Haath system, might close the communication gap between the average person and those who have voice impairments.
C. Finger Detection for Sign Language Recognition
ABSTRACT: A significant scientific challenge for improving communication with hearing-impaired persons is computer identification of sign language. In this work, a quick and effective algorithm for counting the number of anchor point and coordinates that are plotted on the palm and on the fingertips. All these coordinates are connected and traced back to find the necessary actions. Boundary tracing and finger tip detection are the basis for the idea of finger detection. For the model to detect the signs there is no requirement of a specific sensor gloves on the hand similar to 3D movie setup, nor does it require the hand to be properly aligned to the camera. Index Terms: Finger detection, image processing, boundary tracing, sign symbol charts, technological infrastructure for the impaired.
IV. SYSTEM ANALYSIS
A. Existing System
Across the globe, many different languages are spoken and translated. The term "Special People" refers to persons who have difficulties hearing and speaking. "The Dumb" and "The Deaf" people, respectively, have difficulty understanding what the other person is attempting to say. Occasionally individuals will misinterpret these communications using sign language, lip reading, or lip sync. Deaf and dumb persons must rely on visual communication because it is very difficult for them to communicate with a regular person. To solve this issue, I created a project for blind and deaf people using the CNN model, which would translate the image into text format.
Those who are deaf or dumb find it extremely difficult to communicate with others, thus they must rely on some form of visual communication.
B. Proposed System
The methodology used in this research is wholly dependent on the hand gesture's form characteristics. Skin tone and texture are not taken into account as additional methods of hand motion detection since they are so sensitive to changes in lighting and other factors. The method turns each word in the English sentence from the Gestures video into a simple English word. The video processing module's CNN technique produces the findings that match. The Sign Writing Image File is obtained and saved in a folder based on the correct match. This folder provided the Natural Language Generation Module with its input.
C. System Design
D. Software Technologies Required
Any Software Application needs an environment in which it can work and access all the resources for it to deliver quick and accurate results. Mainly Machine Learning Projects are having a heavy demand for a perfectly conditioned environment setup with all the packages installed, all the environment variables set and the dependencies along with their versions mapped.
Major Softwares needed for the Project are:
E. Hardware Requirements
For any efficient software to run we need a proper hardware for which the software is programmed to work in. The basic hardware amenities like Keyboard, mouse, speakers for the voice output, Monitor for displaying the text output in the console and a Physical Webcam to capture the camera feed.
V. IMPLEMENTATION
We use Python, CNN and Keras which are used to detect images.
The Mediapipe Side of Implementation:
A. Algorithm
2. Few Applications of CNN include
a. MRI Imaging
b. Computer Vision applications for self driven cars.
c. Music genre classification
d. Medical diagnosis and transcripts
e. CCTV camera footage detection and analysis for crime reporting.
f. Sign Language Processing
g. Cyber Forensics.
B. Explanation Steps of Implementation
2. Data.pickle File is in the format of .pickle. Generally datasets are used for testing and training of the Data model.
Our Model has images and videos as inputs, we have to feed around 100 images for testing and training the model.We converted the datasets into metadata which is easily readable by the computer and can be accessed fast. This improves the efficiency of the algorithm.
3. The Requirements File is a text document which is present in the Master Folder. It contains all the requirements needed to set up the environment in the project followed by the version description. It enables us to download all the requirements by passing the file as a parameter instead of importing each package individually.
4. The Assets and Variables folders contain all the necessary mappings to the datasets and the necessary variables used to store the values. This is where the architecture of the model is connected to the CNN and the Tensorflow workflows.
5. The Gesture File is a Naming file which maps a particular gesture to the label it is provided with. We must make sure that no changes are made to the Name file because the mappings will be altered and we might get the wrong output.
6. To execute the project we need to Run the “python test.py” command on the Command prompt.
7. This is how the Output is Obtained and the Application is hardcoded in such a way that all the logs of the gestures which have to provided and the predictions it has done are registered in the log of the Command Prompt.
VI. SOFTWARE ENVIRONMENT
A. Implementation Process and Steps for Downloading and Installing Python
Many updates have been done to Python Interpreter recently, after every new update the older versions have started to lose their usability as new versions contain many new features and many new bug fixes.
First of all check with the System Requirements of your computer. Depending on your Operating System and your processor and also how many cores it possesses we need to choose the particular interpreter to download. Our system type is a Windows 64-bit operating system. The most Prominent Version of the python available is Python 3.7.
B. Download the Accurate python Version suitable for your PC
Link: https://www.python.org
2. Download It.
3. Since all the Requirements for our Project are satisfied by the Python 3.7.4 we can download it.
4. The .exe or the .zip file can be downloaded from the website. In case we are downloading the .zip or .rar file we must unzip the compressed files into the local folder and run it as administrator to get the interpreter installed.
After Downloading the Files we must Install Python
a. Find the folder where you downloaded the Python executable file. Right Click on it and select Run as Administrator.
b. Select the Path where we want to place the Python Interpreter Architecture and environment.
c. Click on the Close Button as soon as your setup is over.
Check your Python Version in Command Prompt:
IX. RESULT AND DISCUSSION
Our Model Turned out have an accuracy of about 86.25%. In the training phase since all the inputs were clear and were part of the learning feed, the accuracy of the model with respect to the epoch was high. But in the testing phase we have given inputs with extreme environment conditions like low light, high contrast, hand being placed very far from the camera etc.
But still our model was able to yield decent results and as we proceeded in new iterations our model kept improving. We stopped the iterations when the past three iterations yielded the same accuracy range.
Though our project has Yielded a great accuracy and good results when compared to many other algorithms, there is always scope for improvement. Whoever refers this document can make improvements and henceforth can improvise the algorithm into a less complex one performing the same functionality. I really hope our fellow people with hearing and speaking impairments are benefited with such applications and I also look forward for science to evolve into something beautiful and which help humankind flourish. I thank both my project guide and project coordinator for supporting and recommending me any improvements and features to the algorithm and the application and making this project of ours into a good result. Thanking you V. Rahul S. S
[1] L. Ku, W. Su, P. Yu and S. Wei, \"A real-time portable sign language translation system,\" 2015 IEEE 58th International Midwest Symposium on Circuits and Systems (MWSCAS), Fort Collins, CO, 2015, pp. 1-4, doi: 10.1109/MWSCAS.2015.7282137. [2] Oscar Koller, Hermann Ney, and Richard Bowden. Deep hand: How to train a cnn on 1 million hand images when your data is continu- ous and weakly labelled. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3793–3802, 2016. [3] N. C. Camgoz, S. Hadfield, O. Koller, H. Ney, and R. Bowden, “Neural sign language translation”, in IEEE Conference on computer vision and pattern recognition (CVPR), 2018. [4] M. Mahesh, A. Jayaprakash and M. Geetha, \"Sign language translator for mobile platforms,\" 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Udupi, 2017, pp. 1176-1181, doi: 10.1109/ICACCI.2017.8126001. [5] AlKhuraym,BatoolYahyaetal.“ArabicSignLanguageRecognitionusing Lightweight CNN-based Architecture.” International Journal of Advanced Computer Science and Applications (2022)
Copyright © 2023 Rahul Sesha Sai Vemparala, Malarselvi S, Vara Prasad Koppineedi, Bhagyasri Burada, Lokesh Raj Kavali. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET50284
Publish Date : 2023-04-10
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here