template abstract_book

2
Automatic Language Translation Software For Aiding Communication Between Indian Sign Language And Spoken English Using Labview Yellapu Madhuri * , G.Anitha ** * 2nd year M.Tech, ** Assistant Professor Department of Biomedical Engineering, SRM University, Kattankulathur-603203, Tamilnadu, India www.srmuniv.ac.in Sign Language (SL) is the natural way of communication of speech and/or hearing-impaired people. A sign is a movement of one or both hands, accompanied with facial expression, which corresponds to a specific meaning. This paper presents SIGN LANGUAGE TRANSLATION software for automatic translation of Indian sign language into spoken English and vice versa to assist the communication between speech and/or hearing impaired people and hearing people. It could be used by deaf community as a translator to people that do not understand sign language, avoiding by this way the intervention of an intermediate person for interpretation and allow communication using their natural way of speaking. The proposed software is standalone executable interactive application program developed using LABVIEW software that can be implemented in any standard windows operating laptop, desktop or an IOS mobile phone to operate with the camera, processor and audio device. For sign to speech translation, the one handed SL gestures of the user are captured using camera; vision analysis functions are performed in the operating system and provide corresponding speech output through audio device. For speech to SL translation the speech input of the user is acquired by microphone; speech analysis functions are performed and provide SL gesture picture display of corresponding speech input. The experienced lag time for translation is little because of parallel processing and allows for instantaneous translation from finger and hand movements to speech and speech inputs to SL gestures. This system is trained to translate one handed SL representations of alphabets (A-Z), numbers (1-9) to speech and 165 word phrases to SL gestures The training database of inputs can be easily extended to expand the system applications. The software does not require the user to use any special hand gloves. The results are found to be highly consistent, reproducible, with fairly high precision and accuracy. AIM : To develop a mobile interactive application program for automatic translation of Indian sign language into spoken English and vice-versa to assist the communication between Deaf people and hearing people. The SL translator should be able to translate one handed Indian Sign language finger spelling input of alphabets (A-Z) and numbers (1-9) to spoken English audio output and 165 spoken English word input to Indian Sign language picture display output. OBJECTIVES: To acquire one handed SL finger spelling of alphabets (A to Z) and numbers (1 to 9) to produce spoken English audio output. To acquire spoken English word input to produce Indian Sign language picture display output. To create an executable file to make the software a standalone application. To implement the software and optimize the parameters to improve the accuracy of translation. To minimize hardware requirements and thus expense while achieving high precision of translation. There is a need for monitoring cerebral perfusion MATERIALS Software Tools used: National Instruments LabVIEW and toolkits LABVIEW 2012 version Vision Development Module Vision acquisition module Hardware tools used Laptop inbuilt webcamera- Acer Crystal Eye Laptop inbuilt speaker-Acer eAudio METHOD: The software is a standalone application. To install the file, follow the instructions that appear in the executable installer file. After installing the application, a Graphical user interfacing (GUI) window opens, from which the full application can be used. The GUI has been created to run the entire application from a single window. It has four pages, each page corresponds to a specific application. PAGE 1 gives a detailed demo of the total software usage. PAGE 2 is for speech to sign language translation. When the “start” button is pressed, a command is sent to the Windows 7 inbuilt Speech Recognizer and it opens a mini window at the top. The first time it is started, a tutorial session begins which gives instructions to setup the microphone and recognize the user’s voice input. Configure the speech recognition software. After the initial training, from the next time the program is executed, it starts speech recognition automatically. To train the system for a different user or change the microphone settings, right click on the Speech Recognizer window and select “Start Speech Tutorial”. To stop the speech recognition software say “Stop listening”. To start speech recognition again say “Start Listening”. When the user utters any of the words listed in the “Phrases” it is displayed in the “Command” indicator. A SL gesture picture corresponding to the speech input is displayed in the “Sign” picture indicator. The score of speech input correlation with the trained word is displayed in the “Score” numeric indicator. Use the exit button to exit the application of speech to SL translation. PAGE 3 is for template preparation for sign to speech translation. To execute the template preparation module, press the “Start” button. Choose the camera to acquire images to be used as templates, from the “Camera Name” list. The acquired image is displayed on “Image” picture indicator. If the display image is good to be used for preparing a template, press “Snap frame”. The snapped image is displayed on “Snap Image” picture display. Draw a region of interest to prepare the template and press “Learn”. The image region in the selected portion of the snapped frame is saved to the folder specified for templates. The saved template image is displayed on “Template Image” picture display. Press “Stop” button to stop execution of template preparation module. PAGE 4 is for Sign to speech translation. Press the “Start” button to start the program. Choose the camera to acquire images to be used for pattern matching, from the “Camera Name” list. The captured images are displayed on the “Input Image” picture display. Press the “Match” button to start comparing the acquired input image with the template images in the data base. In each iteration the input image is checked for pattern match with one template. When the input image matches with the template image, the loop halts. The “Match” LED glows and the matched template is displayed on the “Template Image” indicator. The loop iteration count is used for triggering a case structure. Depending on the iteration count value a specific case is selected and gives a string output. Otherwise the loop continues to next iteration where the input image is checked for pattern match with a new template. The information in the string output from case structure is displayed on the “Matched Pattern” alphanumeric indicator. It also initiates the .NET speech synthesizer to give an audio output through the speaker. Figure 1.1 Events involved in hearing Figure 1.2 Speech chain Figure 1.3 Graphical Abstract

Upload: madhuri-yellapu

Post on 16-Jan-2017

340 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Template  abstract_book

Automatic Language Translation Software For Aiding Communication Between Indian Sign Language And Spoken

English Using Labview Yellapu Madhuri*, G.Anitha**

* 2nd year M.Tech, ** Assistant Professor

Department of Biomedical Engineering,

SRM University, Kattankulathur-603203, Tamilnadu, India

www.srmuniv.ac.in

Sign Language (SL) is the natural way of communication of speech and/or

hearing-impaired people. A sign is a movement of one or both hands,

accompanied with facial expression, which corresponds to a specific meaning.

This paper presents SIGN LANGUAGE TRANSLATION software for

automatic translation of Indian sign language into spoken English and vice

versa to assist the communication between speech and/or hearing impaired

people and hearing people. It could be used by deaf community as a translator

to people that do not understand sign language, avoiding by this way the

intervention of an intermediate person for interpretation and allow

communication using their natural way of speaking. The proposed software is

standalone executable interactive application program developed using

LABVIEW software that can be implemented in any standard windows

operating laptop, desktop or an IOS mobile phone to operate with the camera,

processor and audio device. For sign to speech translation, the one handed SL

gestures of the user are captured using camera; vision analysis functions are

performed in the operating system and provide corresponding speech output

through audio device. For speech to SL translation the speech input of the user

is acquired by microphone; speech analysis functions are performed and

provide SL gesture picture display of corresponding speech input. The

experienced lag time for translation is little because of parallel processing and

allows for instantaneous translation from finger and hand movements to speech

and speech inputs to SL gestures. This system is trained to translate one handed

SL representations of alphabets (A-Z), numbers (1-9) to speech and 165 word

phrases to SL gestures The training database of inputs can be easily extended to

expand the system applications. The software does not require the user to use

any special hand gloves. The results are found to be highly consistent,

reproducible, with fairly high precision and accuracy.

AIM :

To develop a mobile interactive application program for automatic

translation of Indian sign language into spoken English and vice-versa to assist

the communication between Deaf people and hearing people. The SL translator

should be able to translate one handed Indian Sign language finger spelling input

of alphabets (A-Z) and numbers (1-9) to spoken English audio output and 165

spoken English word input to Indian Sign language picture display output.

OBJECTIVES:

•To acquire one handed SL finger spelling of alphabets (A to Z) and numbers (1

to 9) to produce spoken English audio output.

•To acquire spoken English word input to produce Indian Sign language picture

display output.

•To create an executable file to make the software a standalone application.

•To implement the software and optimize the parameters to improve the accuracy

of translation.

•To minimize hardware requirements and thus expense while achieving high

precision of translation.

There is a need for monitoring cerebral perfusion

MATERIALS

Software Tools used: National Instruments

LabVIEW and toolkits

•LABVIEW 2012 version

•Vision Development Module

•Vision acquisition module

Hardware tools used

•Laptop inbuilt webcamera- Acer Crystal Eye

•Laptop inbuilt speaker-Acer eAudio

METHOD:

The software is a standalone application. To install the file, follow the

instructions that appear in the executable installer file. After installing the

application, a Graphical user interfacing (GUI) window opens, from which the

full application can be used. The GUI has been created to run the entire

application from a single window. It has four pages, each page corresponds to a

specific application.

PAGE 1 gives a detailed demo of the total software usage.

PAGE 2 is for speech to sign language translation.

When the “start” button is pressed, a command is sent to the Windows 7 inbuilt

Speech Recognizer and it opens a mini window at the top. The first time it is

started, a tutorial session begins which gives instructions to setup the microphone

and recognize the user’s voice input. Configure the speech recognition software.

After the initial training, from the next time the program is executed, it starts

speech recognition automatically. To train the system for a different user or

change the microphone settings, right click on the Speech Recognizer window

and select “Start Speech Tutorial”. To stop the speech recognition software say

“Stop listening”. To start speech recognition again say “Start Listening”. When

the user utters any of the words listed in the “Phrases” it is displayed in the

“Command” indicator. A SL gesture picture corresponding to the speech input is

displayed in the “Sign” picture indicator. The score of speech input correlation

with the trained word is displayed in the “Score” numeric indicator. Use the exit

button to exit the application of speech to SL translation.

PAGE 3 is for template preparation for sign to speech translation.

To execute the template preparation module, press the “Start” button.

Choose the camera to acquire images to be used as templates, from the “Camera

Name” list. The acquired image is displayed on “Image” picture indicator. If the

display image is good to be used for preparing a template, press “Snap frame”.

The snapped image is displayed on “Snap Image” picture display. Draw a region

of interest to prepare the template and press “Learn”. The image region in the

selected portion of the snapped frame is saved to the folder specified for

templates. The saved template image is displayed on “Template Image” picture

display. Press “Stop” button to stop execution of template preparation module.

PAGE 4 is for Sign to speech translation.

Press the “Start” button to start the program. Choose the camera to acquire

images to be used for pattern matching, from the “Camera Name” list. The

captured images are displayed on the “Input Image” picture display. Press the

“Match” button to start comparing the acquired input image with the template

images in the data base. In each iteration the input image is checked for pattern

match with one template. When the input image matches with the template image,

the loop halts. The “Match” LED glows and the matched template is displayed on

the “Template Image” indicator. The loop iteration count is used for triggering a

case structure. Depending on the iteration count value a specific case is selected

and gives a string output. Otherwise the loop continues to next iteration where the

input image is checked for pattern match with a new template. The information

in the string output from case structure is displayed on the “Matched Pattern”

alphanumeric indicator. It also initiates the .NET speech synthesizer to give an

audio output through the speaker.

Figure 1.1 Events involved in hearing Figure 1.2 Speech chain

Figure 1.3 Graphical Abstract

Page 2: Template  abstract_book

[1]. Yellapu Madhuri, G.Anitha (2013) “VISION-BASED SIGN

LANGUAGE TRANSLATION DEVICE” International Conference on

Information Communication & Embedded systems ICICES 2013 in association

with IEEE, S.A engineering College, Chennai. ISBN No. 978-1-4673-5787-6G.

Tracking Id: 13cse213.

[2]. Yellapu Madhuri, G.Anitha (2013) “Automatic Language Translation

Software for Interpreting Sign Language and Speech in English”, has been

awarded Silver medal in paper presentation Research Day 2013 at SRM

University, Chennai.

[3]. Yellapu Madhuri, G.Anitha (2013) submission entitled "SIGN

LANGUAGE TRANSLATOR" has been assigned the following manuscript.

number: IMAVIS-D-13-00011 by Elsevier Editorial Systems- Image and

Vision Computing journal [email protected].

[4]. Yellapu Madhuri, G.Anitha (2013) submission entitled “VISION-BASED

SIGN LANGUAGE TRANSLATOR” is Accepted for publication

in International Journal of Engineering and Science Invention (IJESI)

www.ijesi.org.Review report of manuscript id: A11023.

[5]. Yellapu Madhuri, G.Anitha (2013) submission entitled “SIGN

LANGUAGE TRANSLATION DEVICE” is Accepted for publication

in The International Journal of Engineering and Science (THE IJES)

www.theijes.com. Review report of manuscript id: 13026.

[6]. Yellapu Madhuri, G.Anitha (2013) submission entitled "Automatic

Language Translation Software for Interpreting Sign Language and Speech in

English" has been assigned a tracking number of NCOMMS-13-02048 by

Nature Communications [email protected].

[1]. Jose L. Hernandez-Rebollar1, Nicholas Kyriakopoulos1, Robert W. Lindeman2 ‘A New Instrumented Approach For Translating American Sign Language Into Sound And Text’, Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition (FGR’04) 0-7695-2122-3/04 $ 20.00 © 2004 IEEE. [2]. K. Abe, H. Saito, S. Ozawa: Virtual 3D Interface System via Hand Motion Recognition From Two Cameras. IEEE Trans. Systems, Man, and Cybernetics, Vol. 32, No. 4, pp. 536–540, July 2002. [3]. Paschaloudi N. Vassilia, Margaritis G. Konstantinos "Listening to deaf': A Greek sign language translator’, 0-7803-9521-2/06/$20.00 §2006IEEE

Name: YELLAPU MADHURI

Reg.No:1651110002

M.Tech (BiomedialEngineering)

Mobile no: 09441571241

E.Mail:[email protected]

In this work, a vision based sign language recognition system using LABVIEW for

automatic sign language translation has been presented. This approach uses the

feature vectors which include whole image frames containing all the aspects of the

sign. This project has investigated the different issues of this new approach to SL

recognition to recognize on the hand sign language alphabets and numbers using

appearance based features which are extracted directly from a video stream recorded

with a conventional camera making recognition system more practical. Although

sign language contains many different aspects from manual and non-manual cues,

the position, the orientation and the configuration or shape of the dominant hand of

the signer conveys a large portion of the information of the signs. Therefore, the

geometric features which are extracted from the signers’ dominant hand, improve

the accuracy of the system to a great degree. This project did not focus on facial

expressions although it is well known that facial expressions convey important part

of sign-languages. The facial expressions can e.g. be extracted by tracking the

signers’ face. Then, the most discriminative features can be selected by employing a

dimensionality reduction method and this cue could also be fused into the

recognition system.

The sign language translator is able to translate alphabets (A-Z) and

numbers (1-9). All the signs can be translated real-time. But signs that are

similar in posture and gesture to another sign can be misinterpreted,

resulting in a decrease in accuracy of the system. The current system has

only been trained on a very small database. Since there will always be

variation in either the signers hand posture or motion trajectory, a larger

database accommodating a larger variety of hand posture for each sign is

required. The speech recognition program requires the user to take up a

tutorial of 10 minutes. During the training, the program learns the accent of

the user for speech recognition. It is observed that, the longer the user used

the program , the higher the accuracy of speech recognition.

This paper presents a novel approach for gesture detection. This approach

has two main steps: i) template preparation, and ii) gesture detection. The

template preparation technique presented here has some important features

for gesture recognition including robustness against slight rotation, small

number of required features and device independence. For gesture detection,

a pattern matching technique is used. The gesture recognition technique

presented here can be used with a variety of front-end input systems such as

vision based input , hand and eye tracking, digital tablet, mouse, and digital

glove. Much previous work has focused on isolated sign language

recognition with clear pauses after each sign. These pauses make it a much

easier problem than continuous recognition without pauses between the

individual signs, because explicit segmentation of a continuous input stream

into the individual signs is very difficult. For this reason, and because of co-

articulation effects, work on isolated recognition often does not generalize

easily to continuous recognition. But the proposed software captures the

input images as an AVI sequence of continuous images. This allows for

continuous input image acquisition without pauses. But each image frame is

processed individually and checked for pattern matching. This technique

overcomes the problem of processing continuous images at the same time

having input stream without pauses.

For Speech to SL translation words of similar pronunciation are

sometimes misinterpreted. This problem can be avoided by clearly

pronouncing the words and with extended training and increasing usage. The

speech recognition technique introduced in this article can be used with a

variety of front-end input systems such as computer and video games,

precision surgery, domestic applications and wearable computers.

Figure 1.4 Block diagram of SL Figure 1.5 Block diagram of Speech

to speech translation to sign language translation Figure 1.6 PAGE 3-GUI of template

preparation

Figure 1.8 PAGE 2-GUI of Speech to

SL translation Figure 1.9 GUI of windows

speech recognition tutorial

Figure 1.7 PAGE 4-GUI of SL to

speech translation

Figure 1.10 Database of SL finger spelling Alphabets and Numbers