Sign language is a way of communicating using hand gestures, movements, and facial expressions, instead of spoken words. It is the medium of communication used by people who are deaf or have hearing impairments to exchange information between their own community and with normal people. In order to bridge the communication gap between people with hearing and speaking disabilities and people who do not use sign language, a lot of research work using machine learning algorithms has been done. Hence, Sign language translator came into picture. Sign Language Translators are generally used to interpret signs and gestures from deaf and hard hearing people and convert them into text.
Introduction
I. INTRODUCTION
Sign language is a visual form of communication that uses body movements and facial expressions to convey meaning between people. For communication with their social environment, Deaf people only use sign language, a non-verbal language. It is based on visual cues sent by the hands, eyes, and face. In sign language, the gestures or symbols are arranged linguistically. It is a complex fusion of nonverbal cues like body language, facial expressions, and hand gestures that convey ideas or thoughts without using words.
There are many spoken languages in the world, and they are all distinctive from one another in some ways. Similar to spoken languages, there are numerous sign languages that use various hand gestures and visual cues. Some of them include Indian Sign Language (ISL), British Sign Language (BSL), French Sign Language (LSF), Pakistani Sign Language (PSL), American Sign Language (ASL), and British Sign Language (BSL). The framework for sign language translation enables deaf, dumb, and speech-impaired people to interact with hearing people using sign language. This results in the elimination of the middleman, who typically serves as a translator. Text to Sign Conversion and Sign to Text Conversion are examples of conversion modules.
A. Motivation
Sign language is used by people who have trouble speaking to interact with others. The traditional approach to gesture recognition involves tracking hand gestures with a camera-based system. Comparatively speaking, the camera-based system is less user-friendly because it would be cumbersome to transport. Additionally, using it in crowded spaces would not be practical because it would pick up multiple gestures from various people who are in its field of view. Separating such offensive gestures is time-consuming and not a practical solution. Basically, we will use the device's built-in camera to capture images, perform vision analysis, operate the operating system, and output speech using the built-in audio device.
II. RELATED WORK
A. Sign Language Translator
The author of the paper [1] has outlined all the methods that can be used to create a standalone Sign Language translator that is installed on a Raspberry Pi and can convert dynamic fingerspells, predict the words that go with them, and construct sentences using the Hand-mesh model that is included in the Mediapipe framework. The generated sentence will also be rendered as audio. Additionally, we have used the face-mesh model, which is also found in MediaPipe, to recognise emotions. We have also included a technique that can recognise text embedded in images on boards, flyers, and other surfaces. Using Google's text-to-speech API, we were able to successfully translate this recognised text into the chosen regional language
B. Sign Language Interpretation
In the paper [2], research was initiated with a number of speeches to text experiments to gauge deaf people's communication abilities and better comprehend their day-to-day issues. The project's main objective was to create a deaf person's communication aid that could be incorporated into a mobile phone. In order to communicate with deaf users, this system displayed a face that was only partially animated. They have many applications and are very helpful
C. Sign Language Machine Translation
[3] The literature review that is part of this paper is split into two sections: the traditional SLT and the neural SLT, which has recently taken over the research scene in a similar way that neural architectures have in the Natural Language Processing (NLP) sector. Transformer layers in particular, along with encoder-decoder neural architectures, have become the industry standard for handling this task. They also provide the opportunity to develop multilingual systems, though this is uncommon for SLs. Additionally, datasets are extremely scarce, and this is especially true for SLT since it is very expensive to annotate SL videos with spoken language text translations. Additionally, this makes it difficult for neural models to learn.
D. Sign Language and Gesture Recognition System
This paper [4] discusses various algorithms and methods that can be applied to identify hand gestures and sign language used by various deaf and dumb people. A more proficient and intuitive tool for human-computer interaction is the hand gesture recognition system. Applications range from sign language interpretation to virtual prototyping to medical education. Sign Language is one of the means of communication for the deaf and dumb people. The aforementioned analysis shows that the field of hand gesture recognition has advanced significantly thanks to vision-based hand gesture recognition. C, C++, and Java are the programming languages used to implement the gesture recognition system. to streamline the process, particularly for image processing operations are needed, MATLAB with image processing toolbox is used.
E. Sign Language Translation Systems for Hearing/Speech Impaired People
[5]The author has examined different methods and tools used for sign language translation as well as the kinds of datasets used. The majority of the systems that have already been developed are in American or British Sign Language, but the survey's primary focus was on Indian Sign Language. Systems for translation in languages like Chinese Sign Language and Russian Sign Language have also been noted. The majority of translation systems only translated one way, either from sign language to text-to-speech or from speech to sign language generation. A full two-way Indian Sign Language sign language system, however, was not seen. Existing systems are either one-way translators or domain-specific. They don't really help many people who are generally deaf or hard of hearing.They want a system that can provide them with a variety of features, such as real-time translation between sign languages. Additionally, the system should be general rather than domain-specific. This would enable them to more effectively and profoundly express their thoughts and feelings. As a result, a two-way Indian Sign Language translation system is required to completely close the communication gap between hearing- and speech-impaired individuals and other people, who could then fit into the majority of the proposed classification.
III. LITURATURE SURVEY
Sr. No
Title
Author/Year of
Publication
Strength
Weakness
1
Literature Survey: Sign Language Translator
Sujay R, Somashekar
M, Aruna Rao B P,
May 2022
Use of MediaPipe module Used Raspberry Pi
Reduction is unstable if image contains Noise
2
Sign Language Interpretation
Prof. D.S.Shingate,
Rutika Bajaj, An-
shu Singh, Gayatri
Walzade, Yogita
Bhavar, Oct 2019
Ready to use Solution Easy to Integrate and works on cross-platform: Open-Sourced
No weaknesses
3
A survey on Sign Language
machine translation
Adri ?an N ?u ?nez Marcos, Olatz Perez-de-Vi ?naspre,
Gorka Labaka, 2018
Faster and more Accurate than other OCR Engines
Need to provide good quality images (Lighting Conditions, Contrast and brightness)
4
Survey on Sign Language and Gesture Recognition System
Hema B N, Sania
Anjum, Umme Hani,
Vanaja P, Akshatha
M, Mar 2019
Easy to use and integrate
Only estimates pose of the human body excludes facial expression
5
Sign Language Translation Systems
for Hearing/Speech Impaired People: A Review
Yuvraj Grover, Riya
A Sharma, 2021
2D features and then optimizes model parameters to fit the features
complexity of the run-time is more
6
Literature
Survey On
Hand Gesture
Techniques For
Sign Language
Recognition
Ms Kamal Preet
Kour, Dr. (Mrs) Lini
Mathew, Aug 2017
The architecture used in the model is simple
No weaknesses
7
Sign Language
Recognition
Karan Bhavsar,
Raj Ghatiya, Aarti
Gohil, Devanshi
Thakkar, Bhumi
Shah, 2021
Ready to use Solution
No weaknesses
8
Sign Language
Recognition
For Deaf And
Dumb People
Using Android
Environment
A. Gayathri, Dr.A.
Sasi Kumar, 2017
Faster and more Accurate
The translation is inaccurate and inappropriate sometimes Requires Internet Connectivity
9
Sign Language to Text and Speech Translation in Real Time Using
Convolutional Neural Network
Ankit Ojha, Ayush
Pandey, Shubham
Maurya, 2020
Easy to use and integrate
No weaknesses
10
Systematic Literature Review: American Sign Language Translator
Andra Ardiansyaha ,
Brandon Hitoyoshia,
Mario Halima,
Novita Hanafiahb,
Aswin Wibisuryac,
2020
Easy to use and integrate
No weaknesses
Conclusion
This paper addresses the problem of communication with deaf and dumb people. Given an input image, the goal is to extract hand landmark key points and create a data set and give an output in text and speech form. We can say that the accuracy of the individual components of our design is good, however, with a huge scope for improvement.
References
[1] Sujay R, Somashekar M, Aruna Rao B P. “Literature Survey: Sign Language Translator” In IJIRT, Volume 8, Issue 12, May 2022
[2] Sandra E. F. de Avila, Antonio da Luz Jr., Arnaldo de A. Araujo, and Matthieu Cord§ † Computer Science Department — Federal University of Minas Gerais. “VSUMM: An Approach for Automatic Video Summarization and Quantitative Evaluation” In IEEE Conference, 2016
[3] Mrs.Poonam S.Jadhava, Prof. Dipti S. Jadhavb * a Department of Information Technology, Ramrao Adik Institute Of Technology, Navi Mumbai,400706, India. “Video Summarization using Higher Order Color Moments (VSUHCM)”
[4] Ting Yao, Tao Mei, and Yong Rui Microsoft Research, Beijing, China. “Highlight Detection with Pairwise Deep Ranking for First-Person Video Summarization ”
[5] Wei Zhang Faculty of Computer Guangdong University of Technology Guangzhou City, Guangdong Province, China “State Transition-Based for Cooperative Shot Boundary Detection”
[6] Zuzana Cernekov ?a, Ioannis Pitas ? , Senior Member, IEEE, and Christophoros Nikou, Member, IEEE “] Information Theory-Based Shot Cut/Fade Detection and Video Summarization”
[7] Shrikant A. Shinde, Sakshi Bondre, Priyanka Parmar, Khushboo Patil, Pratiksha Yewalkar, “An Social Distance Violation Detector System”, International Journal of Innovative research in Engineering, Vol 3 , Issue 3, PP 191-195, May-June 20222.
[8] Shrikant A. Shinde, Bhavesh Patil, Mrunali Ghate, Poonam Shinare, Ajay Patil, “Skin Burn and Skin Cancer Detection using Image Processing ”, International ResearchJournal of Engineering and Technology, Vol 9, Issue 6, June 2022.
[9] Shrikant A Shinde, Shailaja N Uke, “Transmission Policies for Data Aggregation using Cooperate Node in Wireless Sensor Networks” IJSR Vol 6, Issue 1, Jan 2017
[10] Shrikant A Shinde, Abhilasha V. Biradar, “Genetic Algorithm for Privacy Protected Personalized Web Search” JETIR, Vol 7, Issue 4, Apr 2020