Automatic Emotion Recognizer for Low Functioning Persons...

5
International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014 1434 ISSN: 2278 7798 All Rights Reserved © 2014 IJSETR AbstractAutism spectrum disorders are a group of developmental disabilities can cause significant social communication and behavioral challenges over lifetime. The theory of causation of autism and the other autism spectrum disorder is incomplete but many number of persons suffering from this disorder. People with autism have issues in non-verbal communication particularly low functioning individual with autism. They might be completely non-verbal and cognitively impaired, lack of understanding and responding to non-verbal cues and communication. Several augmentative systems were developed for low functioning people but they act only as a learning environment and often failed to operate in conjunction with real world situations. This system must also require a manual intervention and assessment reports of an individual autistic person. So we proposed a new therapeutic system called automatic emotion recognizer that recognizes Virtual Reality based facial expressions there by finding different kinds of emotions i.e. neutral, smile, anger, sad, surprise of an autistic person. Index Termsautism spectrum disorder, facial expression, low functioning people, virtual reality. I. INTRODUCTION Autism Spectrum Disorder is highly variable neuro developmental disorder which is characterized by great difficulty in communicating with others, abnormal behavior patterns and forming relationships. People with ASD have deficits in three major areas: verbal and non-verbal communication, social awareness and interaction, imaginative play and cognitive inflexibility. Diagnosis of autism is based on mainly behavioral assessment, parental report and clinical report. In some cases, Diagnostician uses tools such as Childhood Autism Rating Scale (CARS) and Autism Diagnostic Observation Scale (ADOS) to measure the severity range of an individual autistic person. . Most of the autistic people face difficulties in communicating with others and also lacking in understanding social cues and convention. They are unable to properly express non-verbal communication and body language. Low Functioning People with Autism might be completely non-verbal, cognitively impaired; they does not communicate even with closest of the family members .These inabilities hinder them from understanding verbal and non-verbal communications, as well as reading human facial expressions effectively. The ability to identify and determine one’s emotions can serve as an empowerment for the field of artificial intelligence and gave rise to smarter, more powerful machines that understands the intention of users. An intelligent machine with emotional awareness can achieve the shortcomings of autistic people. With that emotional awareness, the machine is capable of teaching and guiding autistic people on how to respond appropriately when the person that he or she is communicating with is expressing various emotions [1]. We developed an automatic emotion recognizer for low functioning persons with autism using Desktop VR tool. This tool breaks the dependency of an autistic person as such it does not need any manual intervention and also operate flexibly in conjunction with real world scenarios. II. EARLY INTERVENTON SYSTEM An innovative VR-based facial emotional expression presentation system was developed that allows monitoring of eye gaze and physiological signals related to emotion identification to explore new efficient therapeutic paradigms. The eye tracking and physiological data were analyzed to determine intergroup and intergroup variations of gaze and physiological patterns. Performance data, physiological signals and eye tracking indices indicated that there were differences in the way adolescents with ASD process and recognize emotional faces compared to their typically developing peers. However, VR systems applied in the context of autism therapy focus on performance or explicit user feedback as primary means of evaluation and thus lack adaptability Traditional behavioral intervention is not accessible to the vast majority of ASD population due to lack of trained therapists as well as intervention costs and often failed in identifying complex expressions as well as required more prompts and more time to facial emotional expression understanding tasks. It uses immersive head mounted displays (HMD) as VR tool that rated as heavy and discomfort. The main problem found in our existing system as, classification systems designed to output one emotion label per input utterance may perform poorly if the expressions cannot be well captured by a single emotional label and multiple algorithm need for finding the Human-emotion. Among the fundamental social impairments in the ASD population are challenges in appropriately recognizing and responding to facial expressions. Innovative technology promises alternative paradigm in increasing intervention accessibility that recognizes VR-based facial expressions recognition system in a synchronous manner. Innovative technology promises alternative or assistive therapeutic paradigms in increasing decreasing assessment efforts, increasing intervention accessibility, reducing the cost of treatment, promoting intervention and ultimately skill generalization. We believe that such ability will provide insight to the emotion recognition process of the children with ASD and eventually help in designing new intervention paradigms to address the emotion recognition vulnerabilities. III. DIAGNOSING TOOL: ADOS The Autism Diagnostic Observation Scale (ADOS) is a standardized , semi-structured which is a combination of the Automatic Emotion Recognizer for Low Functioning Persons with Autism in Virtual Reality Environment G.Aparna 1 , S.Srinivasan 2 1 Department of Electronics and Communication, P.B. College of Engineering, Chennai. 2 Professor, Dept of Information and Technology, P.B. College of Engineering, Chennai.

Transcript of Automatic Emotion Recognizer for Low Functioning Persons...

Page 1: Automatic Emotion Recognizer for Low Functioning Persons ...ijsetr.org/wp-content/uploads/2014/05/IJSETR-VOL-3... · eye gaze. The purpose of this study was to conduct a systematic

International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014

1434

ISSN: 2278 – 7798 All Rights Reserved © 2014 IJSETR

Abstract— Autism spectrum disorders are a group of

developmental disabilities can cause significant social

communication and behavioral challenges over lifetime. The

theory of causation of autism and the other autism spectrum

disorder is incomplete but many number of persons suffering

from this disorder. People with autism have issues in non-verbal

communication particularly low functioning individual with

autism. They might be completely non-verbal and cognitively

impaired, lack of understanding and responding to non-verbal

cues and communication. Several augmentative systems were

developed for low functioning people but they act only as a

learning environment and often failed to operate in conjunction

with real world situations. This system must also require a

manual intervention and assessment reports of an individual

autistic person. So we proposed a new therapeutic system called

automatic emotion recognizer that recognizes Virtual Reality

based facial expressions there by finding different kinds of

emotions i.e. neutral, smile, anger, sad, surprise of an autistic

person.

Index Terms— autism spectrum disorder, facial expression, low

functioning people, virtual reality.

I. INTRODUCTION

Autism Spectrum Disorder is highly variable neuro

developmental disorder which is characterized by great difficulty in communicating with others, abnormal behavior

patterns and forming relationships. People with ASD have

deficits in three major areas: verbal and non-verbal

communication, social awareness and interaction,

imaginative play and cognitive inflexibility. Diagnosis of

autism is based on mainly behavioral assessment, parental

report and clinical report. In some cases, Diagnostician uses

tools such as Childhood Autism Rating Scale (CARS) and

Autism Diagnostic Observation Scale (ADOS) to measure

the severity range of an individual autistic person. . Most of

the autistic people face difficulties in communicating with

others and also lacking in understanding social cues and convention. They are unable to properly express non-verbal

communication and body language. Low Functioning People

with Autism might be completely non-verbal, cognitively

impaired; they does not communicate even with closest of the

family members .These inabilities hinder them from

understanding verbal and non-verbal communications, as

well as reading human facial expressions effectively. The

ability to identify and determine one’s emotions can serve as

an empowerment for the field of artificial intelligence and

gave rise to smarter, more powerful machines that

understands the intention of users. An intelligent machine with emotional awareness can achieve the shortcomings of

autistic people. With that emotional awareness, the machine

is capable of teaching and guiding autistic people on how to

respond appropriately when the person that he or she is

communicating with is expressing various emotions [1]. We

developed an automatic emotion recognizer for low

functioning persons with autism using Desktop VR tool. This

tool breaks the dependency of an autistic person as such it

does not need any manual intervention and also operate

flexibly in conjunction with real world scenarios.

II. EARLY INTERVENTON SYSTEM

An innovative VR-based facial emotional expression

presentation system was developed that allows monitoring of

eye gaze and physiological signals related to emotion

identification to explore new efficient therapeutic paradigms.

The eye tracking and physiological data were analyzed to

determine intergroup and intergroup variations of gaze and

physiological patterns. Performance data, physiological

signals and eye tracking indices indicated that there were

differences in the way adolescents with ASD process and recognize emotional faces compared to their typically

developing peers. However, VR systems applied in the

context of autism therapy focus on performance or explicit

user feedback as primary means of evaluation and thus lack

adaptability Traditional behavioral intervention is not

accessible to the vast majority of ASD population due to lack

of trained therapists as well as intervention costs and often

failed in identifying complex expressions as well as required

more prompts and more time to facial emotional expression

understanding tasks. It uses immersive head mounted

displays (HMD) as VR tool that rated as heavy and discomfort. The main problem found in our existing system

as, classification systems designed to output one emotion

label per input utterance may perform poorly if the

expressions cannot be well captured by a single emotional

label and multiple algorithm need for finding the

Human-emotion. Among the fundamental social impairments

in the ASD population are challenges in appropriately

recognizing and responding to facial expressions. Innovative

technology promises alternative paradigm in increasing

intervention accessibility that recognizes VR-based facial

expressions recognition system in a synchronous manner. Innovative technology promises alternative or assistive

therapeutic paradigms in increasing decreasing assessment

efforts, increasing intervention accessibility, reducing the

cost of treatment, promoting intervention and ultimately skill

generalization. We believe that such ability will provide

insight to the emotion recognition process of the children

with ASD and eventually help in designing new intervention

paradigms to address the emotion recognition vulnerabilities.

III. DIAGNOSING TOOL: ADOS

The Autism Diagnostic Observation Scale (ADOS) is a

standardized , semi-structured which is a combination of the

Automatic Emotion Recognizer for Low

Functioning Persons with Autism in Virtual

Reality Environment G.Aparna1, S.Srinivasan2

1Department of Electronics and Communication, P.B. College of Engineering, Chennai. 2Professor, Dept of Information and Technology, P.B. College of Engineering, Chennai.

Page 2: Automatic Emotion Recognizer for Low Functioning Persons ...ijsetr.org/wp-content/uploads/2014/05/IJSETR-VOL-3... · eye gaze. The purpose of this study was to conduct a systematic

International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014

1435

ISSN: 2278 – 7798 All Rights Reserved © 2014 IJSETR

two earlier instruments called ICD-10 and DSM-IV

diagnostic criteria used by professional for diagnosing

autism. It assesses communication, social interaction and

imaginative use of materials for individuals who may have an

autism spectrum disorder. The ADOS consist of four

modules, each based on language and age

I- Preverbal / Single Words

II- Phase Speech

III- Fluent Speech, Child/Adolescent

IV- Fluent Speech, Adolescent/ Adult

Each module contains standard activities and materials that

are presented by examiners in order to elicit behaviors that

have been used for identifying and also diagnosis of an ASD

at that age and language level i.e. eye contact, conversation,

use of speech and language, unusual sensory interests and

shared enjoyment. The aim of the activities is to structure the interactions so that the child (or) adult being assessed is

sufficiently interacted those subjects want to participate

socially. In 2002 lord, rutter, dilavore & risi reported ADOS

manual, intraclasss correlations are as follows: test-retest

reliability from .59 to .73 and interrater reliability from .82 to

.93.cronbach’s alphas for internal consistency were

consistently highest for the communication-social interaction

total score from .91 to .94 and lowest for stereotyped

behaviors and restricted interests score (.63 to .65 for

modules 2 &1 and .47 to .56 for modules 4 and 3). Each

module can be administrated in 30-40 minutes and notes are taken by the examiner during its administration. The overall

ratings are completed immediately after the administration,

which are then used to formulate a diagnosis through the use

of the diagnostic algorithm provided for each module.

Scoring made at the end of the module is similar across

modules with some identical items, but others are relevant

only for the module being used. In this study, their results

indicated substantial interrater and test-retest reliability for

individual items, excellent interrater reliability within

domains and external internal consistency. We considered

only two autistic samples under the age group of 31 and 34

with moderate autism.

IV. SYSTEM DESIGN

A. Non-Verbal Communication

Low functioning individual with autism are completely

non-verbal; they express their feelings and share their own

emotions via facial expression, body postures, gestures and eye gaze. The purpose of this study was to conduct a

systematic examination of nonverbal communicative

behaviors using multiple perspectives (both trained observers

and instructors familiar with children's idiosyncratic

communication) and to explore instructors' responsively to

these subtle but important cues. More specifically, the

objectives in this study were as follows:

Fig .1 Autistic Persons with different kinds of Facial Expression

To compare the frequency of potentially communicative

behaviors’ displayed by a group of students with autism, as

identified by trained observers and familiar instructors; To

identify the communicative functions perceived by familiar

instructors; To determine the proportion of communicative

behaviors to which instructors responded and

determine/identify different kinds of facial expressions and

emotions of an individual who is suffered from ASD.Further,

a more elaborate approach to responsively coding, including a broader range of instructors' responses and sequential

analysis, would provide a richer understanding of social

interactions.

B. Experimental Setup

An experiment was performed in a laboratory with two

rooms separated by one-way glass windows for caregiver

observation. In the inner room, the subject sat in front of the

task computer. The caregivers sat in the outside room. A

therapist was present in the inner room to monitor the

process. The task computer monitor was also routed to the

outer room for caregiver observation during training session.

The session was video recorded for the whole duration of

participation.

C. Architectural Design

Initially an image is captured by web camera; face detection

is done using skin tone detection by finding skin-colored

pixels and regions in an image then converting the image into

the HSV color space and classifies to either skin or non-skin

those results in much stretched skin color cluster image. The

face clipped image takes as input for feature extraction that

reduces dimensionality as well as extracts both transient and intransient features i.e. eyes, eyebrows and mouth to perform

the desired task.

Page 3: Automatic Emotion Recognizer for Low Functioning Persons ...ijsetr.org/wp-content/uploads/2014/05/IJSETR-VOL-3... · eye gaze. The purpose of this study was to conduct a systematic

International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014

1436

ISSN: 2278 – 7798 All Rights Reserved © 2014 IJSETR

Fig. 2 VR-based Automatic Emotion Recognizer System

In this module Local Binary Pattern (LBP) ) exploits

textual nature of human face as well as relationship between

component features via eyes, nose and mouth to detect face

patterns and reduce the dimension of the feature space.. After

applying a textual descriptor, facial features become darker.

For expression analysis, intransient features are selected for

the final emotion-specific feature set. This feature set was reduced using information gain on per emotion class basics

and permit as ranking of the features was implemented using

modified fuzzy C means clustering. Then filtering has been

applied to find the edges of intransient features with the help

Canny’s edge detector. Filtered feature set is analyzed and

classified to maintain trained dataset .Finally, VR-based

facial expressions are recognized there by emotions are

identified for accessing and controlling in the modern

environment.

V. SYSTEM IMPLEMENTATION

When user login has been done in face emotion detection

system, an image of autistic person is taken as input by

clicking browse button

Fig. 3 Human Detection

After that by clicking skin button , a skin detector typically

transforms image in to skin or non- skin pixel and then skin classifier is used to label the pixels for decision boundary to

detect whether it is an human or not.

Fig. 4 Skin Tone Analysis

After recognizing a given image is human then connected button is clicked to apply RGB color space which

is used to isolate the presence of arms, face and gestures as

well as eliminate the illumination condition to best extent.

Though it is a face emotion detection system, remove regions

that are unlikely to represent faces and then face acquisition

is done by examining common features on face.

Fig. 5 Applying RGB color space

Apply Local Binary Pattern by clicking binary button.

Face image can be viewed as texture pattern it provides powerful image representation and feature recognition for

further image processing

Fig.6 Feature Extraction

By clicking a face button, the end result of extraction

task is shown called feature vector that contains both

transient and intransient features.

Camera Skin

Tone Detection

Feature Extraction

Feature Selection

Feature

Filtering

Feature Set

Analysis

Face

Expression

Accessing &

Controlling Process

Page 4: Automatic Emotion Recognizer for Low Functioning Persons ...ijsetr.org/wp-content/uploads/2014/05/IJSETR-VOL-3... · eye gaze. The purpose of this study was to conduct a systematic

International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014

1437

ISSN: 2278 – 7798 All Rights Reserved © 2014 IJSETR

Fig.7 Facial Feature Analysis

After feature extraction, selection helps to reduce the

feature vector by removing irrevalant, redundant and noisy

features. It selects only intransient features by clicking

eye-lip button..

Fig. 8 Feature Selection using Clustering Algorithm

Canny’s filtering is applied to detect edges of intransient

features .these edges are to be maintained in an edge pixel database for further processing.

Fig. 9 Feature filtering using Canny’s edge detector

After feature filtering, trained dataset has been created

for each autistic person according to 5 different kinds of

emotions. In the "Person” table, it stores the name of autistic

people and their index of 5 kinds of emotion. In the “Position” table, for each index, there are 6 control points for

lip curve, 5 control points for left eye curve, 6 control points

for right Bezier curve.

Fig.10 Trained Dataset

By clicking emotion button, Automatic emotion

recognizer system automatically detects frontal faces from

the captured image and codes them with respect to 5

dimensions in real time: neutral, anger, smile, sadness,

surprise. Emotions are analyzed from training dataset based

on classification technique called decision tree.

Fig. 11 Emotion Detection

VI. CONCLUSION

We have developed a VR-based automatic emotion

recognizer system that was able to collect facial features data

while the subjects were involved in an emotion recognition

tasks. Specifically, we developed controllable levels of facial

expressions of emotion based on long standing research

documenting certain universal expression patterns as well as

a desktop virtual reality presentation to avoid issues and sensitivities individuals with ASD.This system must be more

useful for low functioning individual with autism and

flexibly operate in real world scenarios. Mainly, it reduces

assessment efforts such as clinical reports, parental feedback,

psychological testing etc.

ACKNOWLEDGMENT

We thank “Aadhuraa Special School, Kanchipuram” for providing support for this work

Page 5: Automatic Emotion Recognizer for Low Functioning Persons ...ijsetr.org/wp-content/uploads/2014/05/IJSETR-VOL-3... · eye gaze. The purpose of this study was to conduct a systematic

International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 5, May 2014

1438

ISSN: 2278 – 7798 All Rights Reserved © 2014 IJSETR

REFERENCES

[1] Emily mower, “ A framework for automatic human emotion

classification using emotion profiles”,IEEE transactions on audio, speech

and language processing, vol-19, pp 1057-1071, July 2011.

[2] H.C.Vijaylakshmi and S.patil kulkurni, “ face detection for skin tone

images using wavelet and texture features”, TECHNIA – international

journal of computing science and technologies , pp 646-650,jan 2011.

[3] Ihsan ullah,Muhammad hussain,ghulam Muhammad,hatim Aboalsamh ,

George Bebis , and Anwar M. Mirza , “ gender recognition from facial

images with local WLD descriptor” , national plans for science and

technology , pp 417-420 ,2012.

[4] Krishna Mohan Kediri, Abas said, and M.Yuntis, “emotion detection

using sub-image based features through human facial expressions”,

international conference on computer and information science, pp 332-335,

2012.

[5] Maringanti hima bindu, Priya gupta, and U.S.Tiwary, “cognitive model

based emotion recognition from facial expressions for live human computer

interaction”, proceedings of the IEEE symposium on computational in

image and signal processing, pp 351-356, 2007

[6] Ming-chuan hung and don-ling yang, “An efficient fuzzy c means

clustering algorithm” , information science , pp 225-232 ,2001

[7] Ryszard S. choras, “image feature extraction techniques and their

applications for CBIR and biometrics system”, international journal of

biology and biomedical engineering, pp 6-16, 2007

[8] Teik toe, shi-min, siu-yeung and yok-yen, “emotional advisor to help

children with autism in social communication”, intelligent decision

technologies, vol-3, pp 278-283, 2012.

[9] Vasantha.M, DR. V. subbiah bharathi, R.dhamodharan, “medical image

feature extraction, selection and classification”, international journal of

engineering science and technology, vol – 2(6), pp 2071-2076, 2010.

[10] Vineet saini, Rajnish garg, “A comparative analysis on edge detection

techniques used in image processing”, IOSR journal of electronics and

communication engineering (IOSRJECE), vol 1, pp 56-59, June 2012.

.