An Application for Performing Real Time Speech Translation in Mobile Environment
Template abstract_book
1. Automatic Language Translation Software For Aiding Communication Between Indian Sign Language And Spoken
English Using Labview
Yellapu Madhuri*, G.Anitha**
* 2nd year M.Tech, ** Assistant Professor
Department of Biomedical Engineering,
SRM University, Kattankulathur-603203, Tamilnadu, India
www.srmuniv.ac.in
Sign Language (SL) is the natural way of communication of speech and/or
hearing-impaired people. A sign is a movement of one or both hands,
accompanied with facial expression, which corresponds to a specific meaning.
This paper presents SIGN LANGUAGE TRANSLATION software for
automatic translation of Indian sign language into spoken English and vice
versa to assist the communication between speech and/or hearing impaired
people and hearing people. It could be used by deaf community as a translator
to people that do not understand sign language, avoiding by this way the
intervention of an intermediate person for interpretation and allow
communication using their natural way of speaking. The proposed software is
standalone executable interactive application program developed using
LABVIEW software that can be implemented in any standard windows
operating laptop, desktop or an IOS mobile phone to operate with the camera,
processor and audio device. For sign to speech translation, the one handed SL
gestures of the user are captured using camera; vision analysis functions are
performed in the operating system and provide corresponding speech output
through audio device. For speech to SL translation the speech input of the user
is acquired by microphone; speech analysis functions are performed and
provide SL gesture picture display of corresponding speech input. The
experienced lag time for translation is little because of parallel processing and
allows for instantaneous translation from finger and hand movements to speech
and speech inputs to SL gestures. This system is trained to translate one handed
SL representations of alphabets (A-Z), numbers (1-9) to speech and 165 word
phrases to SL gestures The training database of inputs can be easily extended to
expand the system applications. The software does not require the user to use
any special hand gloves. The results are found to be highly consistent,
reproducible, with fairly high precision and accuracy.
AIM :
To develop a mobile interactive application program for automatic
translation of Indian sign language into spoken English and vice-versa to assist
the communication between Deaf people and hearing people. The SL translator
should be able to translate one handed Indian Sign language finger spelling input
of alphabets (A-Z) and numbers (1-9) to spoken English audio output and 165
spoken English word input to Indian Sign language picture display output.
OBJECTIVES:
•To acquire one handed SL finger spelling of alphabets (A to Z) and numbers (1
to 9) to produce spoken English audio output.
•To acquire spoken English word input to produce Indian Sign language picture
display output.
•To create an executable file to make the software a standalone application.
•To implement the software and optimize the parameters to improve the accuracy
of translation.
•To minimize hardware requirements and thus expense while achieving high
precision of translation.
There is a need for monitoring cerebral perfusion
MATERIALS
Software Tools used: National Instruments
LabVIEW and toolkits
•LABVIEW 2012 version
•Vision Development Module
•Vision acquisition module
Hardware tools used
•Laptop inbuilt webcamera- Acer Crystal Eye
•Laptop inbuilt speaker-Acer eAudio
METHOD:
The software is a standalone application. To install the file, follow the
instructions that appear in the executable installer file. After installing the
application, a Graphical user interfacing (GUI) window opens, from which the
full application can be used. The GUI has been created to run the entire
application from a single window. It has four pages, each page corresponds to a
specific application.
PAGE 1 gives a detailed demo of the total software usage.
PAGE 2 is for speech to sign language translation.
When the “start” button is pressed, a command is sent to the Windows 7 inbuilt
Speech Recognizer and it opens a mini window at the top. The first time it is
started, a tutorial session begins which gives instructions to setup the microphone
and recognize the user’s voice input. Configure the speech recognition software.
After the initial training, from the next time the program is executed, it starts
speech recognition automatically. To train the system for a different user or
change the microphone settings, right click on the Speech Recognizer window
and select “Start Speech Tutorial”. To stop the speech recognition software say
“Stop listening”. To start speech recognition again say “Start Listening”. When
the user utters any of the words listed in the “Phrases” it is displayed in the
“Command” indicator. A SL gesture picture corresponding to the speech input is
displayed in the “Sign” picture indicator. The score of speech input correlation
with the trained word is displayed in the “Score” numeric indicator. Use the exit
button to exit the application of speech to SL translation.
PAGE 3 is for template preparation for sign to speech translation.
To execute the template preparation module, press the “Start” button.
Choose the camera to acquire images to be used as templates, from the “Camera
Name” list. The acquired image is displayed on “Image” picture indicator. If the
display image is good to be used for preparing a template, press “Snap frame”.
The snapped image is displayed on “Snap Image” picture display. Draw a region
of interest to prepare the template and press “Learn”. The image region in the
selected portion of the snapped frame is saved to the folder specified for
templates. The saved template image is displayed on “Template Image” picture
display. Press “Stop” button to stop execution of template preparation module.
PAGE 4 is for Sign to speech translation.
Press the “Start” button to start the program. Choose the camera to acquire
images to be used for pattern matching, from the “Camera Name” list. The
captured images are displayed on the “Input Image” picture display. Press the
“Match” button to start comparing the acquired input image with the template
images in the data base. In each iteration the input image is checked for pattern
match with one template. When the input image matches with the template image,
the loop halts. The “Match” LED glows and the matched template is displayed on
the “Template Image” indicator. The loop iteration count is used for triggering a
case structure. Depending on the iteration count value a specific case is selected
and gives a string output. Otherwise the loop continues to next iteration where the
input image is checked for pattern match with a new template. The information
in the string output from case structure is displayed on the “Matched Pattern”
alphanumeric indicator. It also initiates the .NET speech synthesizer to give an
audio output through the speaker.
Figure 1.1 Events involved in hearing Figure 1.2 Speech chain
Figure 1.3 Graphical Abstract