Hand Signs To Audio Converte1
Hand Signs To Audio Converte1
Hand Signs To Audio Converte1
PROJECT SYNOPSIS
OF MAJOR PROJECT
BACHELOR OF TECHNOLOGY
CSE
Anish Thakur
(2002300100024)(15083)
Rahul
(2002300100127)(15105)
Saurabh Agarwal
(2002300100142)(15682)
1. Abstract
2. Introduction
Overall description
Purpose
Motivations and scope
3. Literature Survey
4. Problem Statement
5. Proposed model
6. References
Abstract
The hand gestures are one of the typical methods used in sign language. It is very
difficult for the hearing impaired people to communicate with the world. This
project presents a solution that will not only automatically recognize the hand
gestures but will also convert it into speech and text output so that impaired person
can easily communicate with normal people. A camera attached to computer will
capture images of hand and the contour feature extraction is used to recognize the
hand gestures of the person. Based on the recognized gestures, the recorded
soundtrack will be played.
The goal of the project is to enhance inclusivityby providing an effective bridge
between the deaf and hearing communities through seamless sign language to
speech coversion.
1. INTRODUCTION
Sign Language is the most natural and expressive way for individuals who are deaf
to communicate. Gestures made with the hands and other body parts, including
facial expressions, are included in sign language. It is primarily used by people
who are deaf or dumb. Sign language recognition refers to the process of
converting the user's signs and gestures into text. Each country has its own sign
language. Indians communicate using Indian Sign Language. Other sign languages,
such as ASL (American Sign Language) and BSL (British Sign Language), are
generally single-handed, whereas ISL uses both hands to make signs. A deaf-dumb
person's desire to speak is rarely understood by the average person. As a result, the
deaf-dumb person uses gestures to communicate his or her needs. Dumb people
communicate with us in their own language. In general, they communicate with
others through sign language. They, however, have difficulty communicating with
others who do not understand sign language. Because new technology is being
developed in our generation, we have created a machine learning computer
programming model that translates sign language to text format and would reduce
the communication gap between normal people and deaf people.
1. OpenCV:
Video Input: OpenCV handles capturing live video feed from a camera or
processing pre-recorded videos.
Hand Detection: Use OpenCV to detect and track the hand in the video frames.
This can be achieved using various techniques like background subtraction,
contour analysis, or more advanced methods like using MediaPipe.
2. MediaPipe:
Hand Gesture Recognition: Utilize MediaPipe's hand tracking model to recognize
and interpret hand gestures captured by OpenCV. MediaPipe offers pre-trained
models for hand detection and gesture recognition.
Extracting Hand Features: Process the hand landmarks provided by MediaPipe to
identify specific hand gestures and their corresponding meanings.
3. Tkinter:
User Interface: Create a graphical user interface (GUI) using tkinter for users to
interact with the application. This interface can include buttons, menus, and other
controls for functionalities like starting/stopping the conversion, selecting
audio/video files, etc.
Purpose.
Here about this research is constructing a machine learning model that can predict
hand gesture from a camera & then turn recognised gesture into voice so that non-
Deaf & non-Dumb people may understand what Deaf & Dumb people are saying.
We are using a deep learning Convolution Neural Network to train hand gesture
images, & we are using that trained model to predict those learnt hand motions
from webcam. we used SVM technique in suggested investigation, although
Python SVM is not reliable for distinguishing hand motion.
This project aims to develop a system that can convert hand gestures into text.
project's objective is to add photographs to database, which will match them &
convert them to text. As part about detection process, hands are observed in
motion. method generates text output, reducing communication gap between
humans & deafmutes.
All the signs cannot be expressed in a single image, the system recognizes sign
language exclusively from images to compensate for the limitations of the existing
system, such as image categorization. As a result, we use CNN and RNN to
classify videos. The spatial properties of the hand signs are extracted using CNN.
The CNN model's output will be fed into the RNN model for sequence modelling,
which will determine which sign is shown in the video. The discovered sign will be
translated into text and speech.
III. MOTIVATION
There has been a lot of research into hand sign language gesture recognition in
recent years. The technology used to recognize gestures is listed below
A. Vision-based
4. Problem statement:
The problem statement for developing a hand sign to audio and video converter
using OpenCV, Tkinter, and MediaPipe involves creating a system capable of real-
time hand gesture recognition from a live camera feed, translating these gestures
into textual or audible outputs, and presenting the interpreted signs through a
graphical user interface. The key components and challenges to address include:
7.References
educative.jo/answers/sign-language-translator-using-opencv