What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio ›...

7
What Can Head and Facial Movements Convey about Positive and Negative Affect? Zakia Hammal 1 , Jeffrey F Cohn 1,2 , Carrie Heike 3 , and Matthew L Speltz 4 1 Robotics Institute, Carnegie Mellon University, Pittsburgh, USA 2 Department of Psychology, University of Pittsburgh, Pittsburgh, USA 3 Seattle Children’s Hospital, Seattle, USA 4 University of Washington School of Medicine, Seattle, USA Abstract—We investigated whether the dynamics of head and facial movements apart from specific facial expressions commu- nicate affect in infants. Age-appropriate tasks were used to elicit positive and negative affect in 28 ethnically diverse 12-month- old infants. 3D head and facial movements were tracked from 2D video. Strong effects were found for both head and facial movements. For head movement, angular velocity and angular acceleration of pitch, yaw, and roll were higher during negative relative to positive affect. For facial movement, displacement, velocity, and acceleration also increased during negative relative to positive affect. Our results suggests that the dynamics of head and facial movements communicate affect at ages as young as 12 months. These findings deepen our understanding of emotion communication and provide basis for studying individual differences in emotion in socio-emotional development. I. INTRODUCTION The past five years have witnessed dramatic advances in automatic affect analysis and recognition [31], [23]. The long- term goal of analyzing spontaneous affective behavior in relatively uncontrolled social interaction has become possible through major advances in computer vision in the development of reliable and fully automatic, person-independent methods for the recognition of facial expression in diverse settings [5], [27], [28]. However, until recently, most research in affective computing has emphasized the morphology of facial expression (e.g., holistic expressions and anatomically-based action units (AUs)) to the neglect of related nonverbal displays and their dynamics. When previous research has used temporal information, it typically is limited to adjacent frames within a temporal window to improve the automatic recognition of AUs or holistic facial expressions or the detection of their temporal segments [24], [16] [26], [29] (for a more complete review please see [31] and [23]). This emphasis on holistic expressions and AUs stems in part from Ekman’s pioneering work on basic emotions [6] and his Facial Action Coding System (FACS) [7]. FACS provided guidance on what to decode for emotion recognition, making possible dramatic advances in automated measurement of fa- cial expression and the emerging field of affective computing. Emphasis on static holistic expressions and AUs, however, ignores two critical aspects of affect communication. One, affect communication is multimodal. People do not com- municate solely by the exchange of facial displays. Facial expression is ”packaged” within other non-verbal modalities that may powerfully communicate emotion [1]. Two, affect communication is dynamic. Faces and heads move and the dynamics of that motion matter to the meaning of affective behavior. As but one example, timing systematically varies be- tween spontaneous and posed smiles [?]. Spontaneous smiles follow ballistic timing; whereas posed smiles fail to. Most previous research in automatic affect analysis has considered only adults. With few exceptions [18], [19], affect related behavior has not been considered in infants. However, both theory and data indicate that affect communication, manifested by nonverbal behaviors (such as head and face), play a critical role in infant’s social, emotional, and cognitive development [13], [2], [19], [25]. Differences in how infants respond to positive and negative emotion inductions are pre- dictive of developmental outcomes that include attachment security [3] and behavioral problems [20]. To understand the beginnings of nonverbal communication, we investigated the extent to which non-verbal behavior apart from specific facial expressions communicates emotion in infants at ages as young as 12 months. Additionally, to inform subsequent research into how the dynamics of nonverbal behavior in infants may be related to change with develop- ment in normative and high-risk samples, we included both typically- and atypically developing infants. We asked to what extent the temporal dynamics of head and facial movements communicate information about positive and negative affect. We hypothesized that infants’ head and facial movements during positive and negative affect systematically differ. Automatic analysis of head and facial movement in in- fants is challenging for several reasons. Compared to adults, head shape differs markedly, and they have smoother skin, fewer wrinkles, and faint eyebrows. Occlusion, due to hands in mouth and extreme head movement, is another common problem. To overcome these challenges, we used a newly developed technique that tracks dense feature points and performs 3D alignment from 2D video [14]. This allowed us to ask to what extent head and facial movements communicate infants’ affect. To elicit positive and negative affect, we used two age-appropriate emotion inductions, Bubble Task and Toy- Removal Task [11]. We then used quantitative measurements to compare the spatiotemporal dynamics of head and facial movements during negative and positive affect. We found strong evidence that head and facial movement communicate affect. Velocity and acceleration of head and fa- cial movements were greater during negative affect compared

Transcript of What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio ›...

Page 1: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

What Can Head and Facial Movements Conveyabout Positive and Negative Affect?

Zakia Hammal1, Jeffrey F Cohn1,2, Carrie Heike3, and Matthew L Speltz41 Robotics Institute, Carnegie Mellon University, Pittsburgh, USA

2 Department of Psychology, University of Pittsburgh, Pittsburgh, USA3 Seattle Children’s Hospital, Seattle, USA

4 University of Washington School of Medicine, Seattle, USA

Abstract—We investigated whether the dynamics of head andfacial movements apart from specific facial expressions commu-nicate affect in infants. Age-appropriate tasks were used to elicitpositive and negative affect in 28 ethnically diverse 12-month-old infants. 3D head and facial movements were tracked from2D video. Strong effects were found for both head and facialmovements. For head movement, angular velocity and angularacceleration of pitch, yaw, and roll were higher during negativerelative to positive affect. For facial movement, displacement,velocity, and acceleration also increased during negative relativeto positive affect. Our results suggests that the dynamics ofhead and facial movements communicate affect at ages as youngas 12 months. These findings deepen our understanding ofemotion communication and provide basis for studying individualdifferences in emotion in socio-emotional development.

I. INTRODUCTION

The past five years have witnessed dramatic advances inautomatic affect analysis and recognition [31], [23]. The long-term goal of analyzing spontaneous affective behavior inrelatively uncontrolled social interaction has become possiblethrough major advances in computer vision in the developmentof reliable and fully automatic, person-independent methodsfor the recognition of facial expression in diverse settings[5], [27], [28]. However, until recently, most research inaffective computing has emphasized the morphology of facialexpression (e.g., holistic expressions and anatomically-basedaction units (AUs)) to the neglect of related nonverbal displaysand their dynamics. When previous research has used temporalinformation, it typically is limited to adjacent frames within atemporal window to improve the automatic recognition of AUsor holistic facial expressions or the detection of their temporalsegments [24], [16] [26], [29] (for a more complete reviewplease see [31] and [23]).

This emphasis on holistic expressions and AUs stems inpart from Ekman’s pioneering work on basic emotions [6] andhis Facial Action Coding System (FACS) [7]. FACS providedguidance on what to decode for emotion recognition, makingpossible dramatic advances in automated measurement of fa-cial expression and the emerging field of affective computing.

Emphasis on static holistic expressions and AUs, however,ignores two critical aspects of affect communication. One,affect communication is multimodal. People do not com-municate solely by the exchange of facial displays. Facialexpression is ”packaged” within other non-verbal modalitiesthat may powerfully communicate emotion [1]. Two, affect

communication is dynamic. Faces and heads move and thedynamics of that motion matter to the meaning of affectivebehavior. As but one example, timing systematically varies be-tween spontaneous and posed smiles [?]. Spontaneous smilesfollow ballistic timing; whereas posed smiles fail to.

Most previous research in automatic affect analysis hasconsidered only adults. With few exceptions [18], [19], affectrelated behavior has not been considered in infants. However,both theory and data indicate that affect communication,manifested by nonverbal behaviors (such as head and face),play a critical role in infant’s social, emotional, and cognitivedevelopment [13], [2], [19], [25]. Differences in how infantsrespond to positive and negative emotion inductions are pre-dictive of developmental outcomes that include attachmentsecurity [3] and behavioral problems [20].

To understand the beginnings of nonverbal communication,we investigated the extent to which non-verbal behavior apartfrom specific facial expressions communicates emotion ininfants at ages as young as 12 months. Additionally, to informsubsequent research into how the dynamics of nonverbalbehavior in infants may be related to change with develop-ment in normative and high-risk samples, we included bothtypically- and atypically developing infants. We asked to whatextent the temporal dynamics of head and facial movementscommunicate information about positive and negative affect.We hypothesized that infants’ head and facial movementsduring positive and negative affect systematically differ.

Automatic analysis of head and facial movement in in-fants is challenging for several reasons. Compared to adults,head shape differs markedly, and they have smoother skin,fewer wrinkles, and faint eyebrows. Occlusion, due to handsin mouth and extreme head movement, is another commonproblem. To overcome these challenges, we used a newlydeveloped technique that tracks dense feature points andperforms 3D alignment from 2D video [14]. This allowed usto ask to what extent head and facial movements communicateinfants’ affect. To elicit positive and negative affect, we usedtwo age-appropriate emotion inductions, Bubble Task and Toy-Removal Task [11]. We then used quantitative measurementsto compare the spatiotemporal dynamics of head and facialmovements during negative and positive affect.

We found strong evidence that head and facial movementcommunicate affect. Velocity and acceleration of head and fa-cial movements were greater during negative affect compared

Page 2: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

with positive affect. Observers time varying behavioral ratingsof infant affect were also related to head and facial dynamics.

II. METHODS

A. Participants

Participants were 28 ethnically diverse 12-month-old infants(M = 13.15, SD = 0.52) recruited as part of a multi-site studyinvolving children’s hospitals in Seattle, Chicago, Los Ange-les, and Philadelphia. Participants in the present study wereprimarily from Seattle Children’s Hospital. Two infants wereAfrican-American, 1 Asian-American, 5 Hispanic-American,16 European-American, and 4 Multiracial. Eight were girls.Twelve infants were mildly affected with craniofacial micro-somia (CFM). CFM is congenital condition associated withvarying degrees of facial asymmetry. Case-control differencesbetween CFM and unaffected infants will be a focus of futureresearch as more infants are ascertained. All parents gave theirinformed consent to the procedures.

B. Observational Procedures

Infants were seated in a highchair at a table with anexperimenter and their mother across from them. Theexperimenter sat to the mothers’ left out of camera view andcloser to the table as shown in Fig.1. Two emotion inductionswere administered with each consisting of multiple trials.

1) Emotion Induction: The “Bubble Task” was intendedto elicit positive affect (surprise, amusement, or interest); the“Toy-Removal Task”, was intended to elicit negative affect(frustration, anger, or distress).

Bubble Task: Soap bubbles were blown toward the centerof the table and below camera view (see Fig.3, first row).Before blowing bubbles, the examiner attempted to buildsuspense (e.g., counting 1-2-3; “Ooh, look at the prettybubbles”; “Can you catch one?”; “Where’s that bubblegoing?”) and maintain infants’ engagement (e.g., “allow allbubbles to pop before continuing again”). This procedure wasrepeated three times.

Toy-Removal Task: The examiner showed a car to theinfant and demonstrated how it works by “running” the carbetween her own hands to generate interest (e.g., car is “woundup” by pulling car back a few inches with all 4 wheels on thetable surface) [11]. The toy was then placed within the infants’reach, allowing them to play with the toy for 15 seconds (seeFig. 4, first row). The examiner then gently took back the toyand placed it inside a clear plastic bin for 30 seconds just outof the infant’s reach to elicit frustration or anger (see Fig. 4,rows 2-4). After 30 seconds, the toy was returned to the infant.This procedure was repeated three times.

Neither task required an active response from the infant.Qualitative observations suggest that head pose and handposition appeared to be independent of task. For instance, asshown in Fig. 3 and 4, frontal and non-frontal pose and handplacements were observed in both tasks. The infant reached

toward the bubbles in Fig. 3 and withdrew their hand from thetoy in Fig. 4.

The trials were recorded using a Sony DXC190 compactcamera at 60 frames per second. Infants’ face orientationto the camera was approximately 15◦ from frontal, withconsiderable head movement.

Fig. 1: Observational Procedure.

2) Manual Continuous Ratings: Despite the standardizedemotion inductions, positive and negative affect could occurduring both tasks. To control for this possibility, manual con-tinuous ratings of valence were obtained from four raters. Foreach of 56 video recordings (28 infants * 2 conditions, Bubbleand Toy-Removal Tasks) they independently rated valencefrom the video on a continuous scale (“−100” (maximum ofnegative affect) to “+100” (maximum of positive affect)) usingcustom software [10]. This produced a time series rating foreach rater and each video. To achieve high effective reliability,following [22], the time series were averaged across raters.Effective reliability was evaluated using intraclass correlation[17], [22]. The ICC was 0.92, which indicated high reliability.

Fig.2 shows the emotional maps obtained from the aggre-gated ratings. Continuous ratings of Bubble Tasks are shownon the right, and continuous ratings of Toy-Removal Tasksare shown on the left. In most cases, segments of positive andnegative affect (delimited in green and red, respectively inFig. 2) alternated throughout the interaction regardless of theexperimental tasks. In a few cases, infants never progressedbeyond negative affect even in the nominally positive BubbleTask (see rows 4 and 6 from the top in Fig. 2). This patternhighlights the importance of using independent criteria toidentify episodes of positive and negative affect instead ofrelying upon the intended goal of each condition.

Overall, infants were more positive in the Bubble Task andmore negative in the Toy-Removal Task. The ratio of positiveto negative affect was significantly higher in the Bubble Taskcompared to the Toy-Removal Task (t = 7.1, df = 27, p <=0.001, see green segments in Fig. 2 left), and the ratio ofnegative to positive affect was significantly higher in the Toy-Removal Task compared to the Bubble Task (t = 7.1, df =27, p <= 0.001, see red segments in Fig. 4 right). Moreover,

Page 3: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

Fig. 2: Emotional maps obtained by the aggregated ratings. Each row corresponds to one infant. Columns correspond to the aggregatedcontinuous rating across frames. Green and red delimit positive and negative affect, respectively

the differences between the selected affective segments werespecific to valence (i.e., positive and negative affect) ratherthan the intensity. The absolute intensity of behavioral ratingsdid not differ between segments of positive and negative affect(p > 0.1).

C. Automatic Tracking of Facial Landmarks and Head Ori-entation

A recently developed 3D face tracker was used to trackthe registered 3D coordinates from 2D video of 49 faciallandmarks, or fiducial points, and the 3 degrees of rigid headmovements (i.e., pitch, yaw, and roll, see Fig. 3 and Fig. 4)[14]. The method is generic; person-specific training is notrequired. More specifically, the tracker uses a combined 3Dsupervised descent method [30], where the shape model isdefined by a 3D mesh and the 3D vertex locations of themesh. The tracker registers a dense parameterized shape modelto an image such that its landmarks correspond to consistentlocations on the face (for details, please see [14]).

To evaluate the concurrent validity of the tracker for headpose tracking, it was compared with the CSIRO cylinder-based 3D head tracker [4] on 16 randomly selected videosof Bubble and Toy-Removal Tasks from 8 infants. Concurrentcorrelations between the two methods for pitch, yaw, and rollwere 0.74, 0.93, and 0.92, respectively. Examples of the trackerperformance are shown in Fig. 3 and Fig. 4, respectively.

III. DATA REDUCTION

In the following we first describe the results of the automatictracking and the manual validation of the tracked data. Wethen describe how the measures of displacement, velocity, andacceleration were computed for both head and facial landmarkmovements.

A. Tracking Results

For each video frame, the tracker output the 3D coordinatesof 49 fiducial points and 3 degrees of freedom of rigid headmovements (pitch, yaw, and roll) or a failure message whena frame could not be tracked. To rule out poorly trackedframes, we visually reviewed tracking results overlaid on thecorresponding video frames. Any frames for which the headpose or feature tracking was visually flawed were removedfrom further analysis (as shown in Fig. 3 and Fig. 4). Framesthat could not be tracked or failed visual review were notanalyzed. Failures were due primarily to self occlusions (e.g.,hand-in-mouth) and extreme head turn (out of frame). For eachinfant the percentage of well tracked frames was lower forToy-Removal Task than for Bubble Task (see Table I).

TABLE I: Proportion of Valid Tracking during Bubble and Toy-Removal Tasks.

Bubble Task Toy-Removal Task

Infants 87 72Note: t = 4.50, p <= 0.01, d f = 27.

B. Measurement of Head Movement

Head angles in the horizontal, vertical, and lateral directionswere selected to measure head movement. These directionscorrespond to head nods (i.e. pitch), head turns (i.e. yaw),and lateral head inclinations (i.e. roll), respectively (see blue,green, and red rows in Fig. 3 and Fig. 4). Head angles wereconverted into angular displacement, angular velocity, andangular acceleration. For pitch, yaw, and roll, angular displace-ment was computed by subtracting the overall mean head anglefrom each observed head angle within each valid segment(i.e., consecutive valid frames). Similarly, angular velocity andangular acceleration for pitch, yaw, and roll, were computed

Page 4: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

Fig. 3: Examples of tracking results (head orientation pitch (green), yaw(blue), and roll (red) and the 49 fiducial points) during Bubble Task.

Fig. 4: Examples of tracking results (head orientation pitch (green), yaw(blue), and roll (red) and the 49 fiducial points) during a Toy-RemovalTask.

as the derivative of angular displacement and angular velocity,respectively. Angular velocity and angular acceleration allowmeasurements of changes in head movement from one frameto the next.

The root mean square (RMS) was then used to measurethe magnitude of variation of the angular displacement, theangular velocity, and angular acceleration. The RMS value ofthe angular displacement, the angular velocity, and the angularacceleration were computed as the square root of the meanvalue of the squared values of the quantity taken over eachconsecutive valid segment for each condition. The RMS of thehorizontal, vertical, and lateral angular displacement, angularvelocity, and angular acceleration were then calculated foreach consecutive valid segment, for each condition, and foreach infant separately. RMSs of head movement as used inthe analyses refer to the mean of the RMSs of consecutivevalid segments.

C. Measurement of Facial Movement

The movement of the 49 detected 3D fiducial points (seeFig. 3 and Fig. 4) was selected to measure facial move-ment. The movement of the 49 fiducial points correspondto the movement (without rigid head movements) of thecorresponding facial features such the eyes, eyebrows, andmouth. First, the mean position in the face of each detectedfiducial point within all valid segments (i.e., consecutive validframes) was computed. The displacement of each one of the49 fiducial points was computed by measuring the Euclideandistance between its current position in the face and its mean

position in the face. The velocity of movement of the 49detected fiducial points was computed as the derivative of thecorresponding displacement, measuring the speed of changeof the fiducial points position on the face from one frame tothe next. Similarly, angular acceleration of each one of the 49detected fiducial points was computed as the derivative of thecorresponding velocities.

The RMS was then used to measure the magnitude ofvariation of the fiducial points displacement, velocity, and ac-celeration, respectively. Similarly to head movement, the RMSvalues of the fiducial points displacement, the fiducial pointsvelocity, and the fiducial points acceleration were computedfor each consecutive valid segment for each condition. RMSsof facial movement as used for analyses refer to the meanof the RMSs of consecutive valid segments. For simplicity, inthe following sections the fiducial points displacement, fiducialpoints velocity, and fiducial points acceleration are referred toas facial displacement, facial velocity, and facial acceleration,respectively.

Because the movements of individual points were highlycorrelated, principal components analysis (PCA) was used toreduce the number of parameters (a compressed representa-tion of the 49 RMSs of facial displacement, velocity, andacceleration , respectively). The first principal components ofdisplacement, velocity, and acceleration accounted for 60%,70%, and 70% of the respective variance and were used foranalysis.

Page 5: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

IV. DATA ANALYSIS AND RESULTS

Because of the repeated-measures nature of the data,repeated-measures analyses of variance (ANOVAs) were usedto test for differences in head and facial movements betweenpositive and negative affect. Separate ANOVAs were used foreach measure (e.g., RMS of angular displacement of headpitch). Student’s paired t-tests were used for post-hoc analysesfollowing significant ANOVAs.

A. Infants’ Head Movement During Positive and NegativeAffect

For RMS angular displacement of pitch, yaw, and roll therewas no effect for condition (F1,52 = 0.39, p = 0.53, F1,52 =1.20, p = 0.27, and F1,52 = 0.01, p = 0.94, respectively)1.

For RMS angular velocity of pitch, yaw, and roll, therewere significant effects for condition (F1,52 = 5.66, p = 0.02,and F1,52 = 11.53, p = 0.001, and F1,52 = 4.69, p = 0.03,respectively).

Similarly, for RMS angular acceleration of pitch and yaw,there were significant effects for condition (F1,52 = 4.30, p =0.04, and F1,52 = 4.90, p = 0.03, respectively), and a marginaleffect for RMS angular acceleration of roll (F1,52 = 3.32,p = 0.07). RMS angular velocities and angular accelerationsincreased in negative affect compared with positive affect (seeTable II).

B. Infants’ Facial Movement During Positive and NegativeAffect

There was a marginal effect for RMS facial displace-ment (F1,52 = 3.23, p = 0.07). RMS facial movements weremarginally higher in negative affect compared with positiveaffect (Table III).

For RMS facial velocity and RMS facial acceleration therewere significant effects for condition (F1,52 = 3.98, p = 0.05and F1,52 = 4.79, p = 0.03, respectively).

C. Correlation Between Head and Facial Movements andCorrelation between Head and Facial Movements and Con-tinuous Behavioral Ratings

Because both head and facial movements for most measureswere faster during negative than positive affect, we wished toinvestigate whether head and facial movements were them-selves correlated and to investigate the correlation betweenhead and facial movements and the behavioral ratings. To re-duce the number of facial movement parameters to a compactrepresentation, PCA was used to reduce the 49 fiducial pointtime series to three time series components that accounted for atotal of 70% of the variance (see Comp-1, Comp-2, and Comp-3 in Table IV and Table V). We then computed the correlationsbetween the continuous behavioral ratings and the infants’time series of angular displacement of pitch, yaw and roll, andfacial displacement (measured using the compressed 3 times

1F1,52 corresponds to the F statistic and its associated degrees of freedom.The first number refers to the degrees of freedom for the specific effect andthe second number to the degrees of freedom of error.

series Comp-1, Comp-2, and Comp-3) during Bubble and Toy-Removal Tasks. The obtained correlations are reported in TableIV and Table V. A one-sample t-test tested the null hypothesisthat the mean correlations did not significantly differ from apopulation mean of zero. The null hypotheses were all rejectedat p <= 0.01.

Head angular displacement and facial displacement wereall moderately correlated with continuous behavioral ratingsof positive and negative affect (see Table V) and with eachother (see Table IV). To test whether the behavioral ratingsof positive and negative affect were more correlated withfacial displacement or with head angular displacement, weused a paired t-test to test the null hypothesis that therewere no differences between correlations. For the BubbleTask, the mean correlations between the behavioral ratingsand facial displacement (using Comp-1) were higher thanbetween the behavioral ratings and head angular displacementfor pitch, yaw, and roll (t = −2.34, p <= 0.02, t = −2.48,p <= 0.02, and t = −3.37, p <= 0.002, d f = 27, for pitch,yaw, and roll, respectively). For the Toy-Removal Task, themean correlations between the behavioral ratings and facialdisplacement (using Comp-1) were higher than between thebehavioral ratings and head angular displacement for yaw,and roll (t = −2.60, p <= 0.01 and t = −2.85, p <= 0.008,d f = 27, for yaw and roll, respectively). Effect sizes of thismagnitude are considered large in behavioral science. Nodifferences were found for pitch (t =−1.21, p > 0.1).

Overall, both head angular displacement and facial displace-ment accounted for significant variation in the continuousratings of infants’ affect; and facial displacement accountedfor more of the variance of the continuous ratings.

V. DISCUSSION AND CONCLUSION

We tested the hypothesis that head and facial movements ininfants vary between positive and negative affect. We foundstriking differences in head and facial dynamics betweenpositive and negative affect. Velocity and acceleration for bothhead and facial movements were greater during negative thanpositive affect.

Were these effects specific to the affective differences be-tween conditions? Alternative explanations might be consid-ered. One is whether the differences were specific to intensityrather than valence. These two dimensions are fundamentalto emotion and affective phenomena according to dimensionalmodels (e.g., the ”circumplex” model of Russell and Bullock).If the intensity of observer ratings were greater during oneor the other condition, this would suggest that the valencedifferences were confounded by intensity. To evaluate thispossibility, we compared observer ratings of intensity betweenconditions. They failed to vary. This negative finding suggeststhat intensity of response was not responsible for the sys-tematic variation we observed between conditions. They werespecific to valence of affect.

Another possible alternative explanation could be demandcharacteristics of the tasks. Toy-removal elicits reaching forthe toy; bubble task elicits head motion toward the bubbles.

Page 6: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

TABLE II: Post-hoc paired t-tests (following significant ANOVAs) for pitch and roll RMS angular displacements, RMS angular velocities,and RMS angular accelerations.

Positive Negative Differences by paired t-test

Angular Velocity M (SD) M (SD) t df p

Pitch 0.68 (0.13) 0.79 (0.19) -3.31 25 0.002Yaw 0.60 (0.11) 0.73 (0.16) -4.29 25 < 0.001Roll 0.33 (0.08) 0.40 (0.14) -2.61 25 0.014

Angular Acceleration

Pitch 0.91 (0.18) 1.03 (0.24) -2.84 25 0.01Yaw 0.78 (0.13) 0.89 (0.22) -2.73 25 0.01Roll 0.41 (0.10) 0.48 (0.18) -1.86 25 0.07

Note: M: Mean of RMSs, t: t-ratio, d f : degrees of freedom,p: probability.

TABLE III: Post-hoc paired t-tests (following significant ANOVAs) for RMS facial displacements, RMS facial velocities, and RMS facialaccelerations.

Positive Negative Differences by paired t-test

M (SD) M (SD) t df p

Facial displacement -0.94 (3.43) 0.87 (3.98) -2.05 25 0.05

Facial Velocity -0.12 (0.35) 0.11 (0.50) -2.61 25 0.01

Facial Acceleration -0.17 (0.44) 0.16 (0.65) -2.61 25 0.01Note: M: Mean of RMSs, t: t-ratio, d f : degrees of freedom,

p: probability.

To consider whether these characteristics may have beena factor, we compared the amount of movement betweenconditions. Displacement of the head failed to vary betweentasks. The lack of differences in head displacement suggeststhat demand characteristics did not account for the differencesin acceleration and velocity we observed.

The obtained results are consistent with prior work inadult psychiatric populations that found that head movementsystematically varies between affective states. For instance,our findings are consistent with a recent study that foundgreater velocity of head motion in intimate couples duringepisodes of high versus low conflict [12]. They are consistentas well with the finding that velocity of head movement wasinversely correlated with depression severity. Previous studieshave found that depression is marked by reductions in generalfacial expressiveness [21] and head movement [8], [15], [9].Together, these findings suggest that head and facial motionare lowest during depressed affect, increase during neutralto positive affect, and increase yet again during conflict andnegative affect. Future research is needed to investigate morefully the relation between affect and the dynamics of head andfacial movement.

Finally, some infants had mild CFM. While CFM couldpotentially have moderated the effects we observed, too fewinfants with CFM have yet been enrolled to evaluate this hy-pothesis. Also, CFM was relatively mild. For these reasons, wefocus on the relation between head movement and expression

independent of CFM status. Once study enrollment is furtheradvanced, the possible role of CFM as a moderator can beexamined.

We also examined the relative importance of infants’ headand facial displacement in the behavioral ratings independentlyof specific facial expressions and AUs. The dynamics ofboth head and facial movement accounted for the variancein behavioral ratings of infants’ affect. The next step will beto evaluate how head and facial movement temporally covarywith specific facial expression and AU.

In summary, we found strong evidence that head and facialmovements vary markedly with affect. Our results suggestthat the dynamics of head and facial movements togethercommunicate affect independently of static facial expressions.

ACKNOWLEDGMENT

Research reported in this publication was supported in partby the National Institute of Dental and Craniofacial Researchof the National Institutes of Health under Award NumberDE022438. The content is solely the responsibility of theauthors and does not necessarily represent the official viewsof the National Institutes of Health.

REFERENCES

[1] Beebe, B., and Gerstman, L. The ”packaging” of maternal stimulation inrelation to infant facial-visual engagement. A case study at four months.Merrill-Palmer Quarterly, 26,(4), pp. 321-339, 1980

Page 7: What Can Head and Facial Movements Convey about Positive ... › ~jeffcohn › biblio › ACII2015.pdf · for the recognition of facial expression in diverse settings [5], [27], [28].

TABLE IV: Mean correlations between pitch, yaw and roll angular displacement and facial displacement.

Bubble Task Toy-Removal Task

Comp-1 Comp-2 Comp-3 Comp-1 Comp-2 Comp-3

Pitch 0.23 0.22 0.14 0.21 0.27 0.18

Yaw 0.20 0.15 0.11 0.17 0.17 0.14

Roll 0.17 0.14 0.12 0.11 0.11 0.14Note: All correlations differed significantly from 0, p ≤ 0.01

TABLE V: Mean correlations between pitch, yaw and roll angular displacement and behavioral ratings and between facial displacementand behavioral ratings.

Bubble Task Toy-Removal Task

Pitch Yaw Roll Pitch Yaw Roll

0.23 0.21 0.18 0.23 0.17 0.17

Comp-1 Comp-2 Comp-3 Comp-1 Comp-2 Comp-3

0.34 0.21 0.27 0.29 0.19 0.28Note: All correlations differed significantly from 0, p ≤ 0.01

[2] J.J. Campos, K. Caplowitz, M.E. Lamb, H.H. Goldsmith, C. Stenberg.Socioemotional development. Haith MM, Campos JJ, eds. Handbook ofchild psychology, II(4), New York: Wiley, pp 1373-1396, 1983.

[3] Cohn, J. F., Campbell, S. B., and Ross, S. Infant response in the still-face paradigm at 6 months predicts avoidant and secure attachment at12 months. Development and Psychopathology, 3, 367-376,1991.

[4] M. Cox, J. Nuevo-Chiquero, J. M. Saragih, and S. Lucey, CSIRO FaceAnalysis SDK, ed. Brisbane, Australia, 10th IEEE Int. Conf. Autom.Face Gesture Recog, Shangai, China, Apr. 2013.

[5] Cohn, J.F., and De la Torre, F. Automated face analysis for affectivecomputing. In R. A. Calvo, S. K. D’Mello, J. Gratch and A. Kappas(Eds.), Handbook of affective computing, pp. 131-150, New York, NY:Oxford, 2015.

[6] Ekman, P. Universals and Cultural Differences in Facial Expressions ofEmotions. In Cole, J. (Ed.), Nebraska Symposium on Motivation pp.207-282. Lincoln, NB: University of Nebraska Press, 1972.

[7] Ekman, P., Friesen, W. V., and Hager, J. C. (Eds.). Facial ActionCoding System [E-book]. Salt Lake City, UT: Research Nexus.Ekman,P., Friesen, W. V., and OSullivan, M. (1988). Smiles when lying. Journalof Personality and Social Psychology, 54, pp. 414-420, 2002

[8] H.U. Fisch, S. Frey, H.P. Hirsbrunner. Analyzing nonverbal behavior indepression. J. Abnorm. Psychol., 92, pp. 307318, 1983

[9] J. Girard, J.F. Cohn, M. H. Mahoor, S.M. Mavadati, Z. Hammal, and D.Rosenwald. Nonverbal social withdrawal in depression: Evidence frommanual and automatic analyses. Image and Vision Computing Journal,2014.

[10] J. Girard. CARMA: Software for continuous affect rating and mediaannotation. Journal of Open Research Software, 2014.

[11] H.H. Goldsmith, M.K. Rothbart. The laboratory temperament assess-ment battery. Eugene, OR: University of Oregon, 1999.

[12] Z. Hammal, J.F. Cohn, and D.T. George. Interpersonal Coordination ofHead Motion in Distressed Couples. IEEE Transactions on AffectiveComputing, 5(2), pp.155-167, 2014

[13] C. Izard, E. Woodburn, K. Finlon, S. Krauthamer-Ewing, S. Grossman,A. Seidenfeld. Emotion knowledge, emotion utilization, and emotionregulation. Emotion Review, 3(44), 2011.

[14] LA Jeni, JF Cohn, T Kanade. Dense 3D Face Alignment from 2D Videosin Real-Time. Proc. IEEE Int’l Conf. Face and Gesture Recognition,2015.

[15] J. Joshi, R. Goecke, G. Parker, M. Breakspear. Can body expressionscontribute to automatic depression analysis?, Proc. IEEE Int’l Conf. Faceand Gesture Recognition, Shangai, China, 2013.

[16] S. Koelstra, M. Pantic, I. Patras. A dynamic texture based approach torecognition of facial actions and their temporal models IEEE Transac-tions on PAMI, 32 (11), pp. 19401954, 2010

[17] K.O. McGraw, S.P. Wong. Forming inferences about some intraclasscorrelation coefficients. Psychological Methods, 1, 30-46, 1996.

[18] D.S. Messinger, M. H. Mahoor, S.M. Chow, and J.F. Cohn. Automatedmeasurement of facial expression in infant-mother interaction: A pilotstudy. Inter. J. of Computer Vision, 14(3), pp. 285-305, 2009.

[19] D.S. Messinger, W. Mattson, M.H. Mohammad, and J.F. Cohn. Theeyes have it: Making positive expressions more positive and negativeexpressions more negative, Emotion, 12, pp. 430-436, 2012.

[20] Moore, G. A., Cohn, J. F., and Campbell, S. B. Infant affective responsesto mother’s still-face at 6 months differentially predict externalizing andinternalizing behaviors at 18 months. Developmental Psychology, 37,pp. 706-714, 2001.

[21] B. Renneberg, K. Heyn, R. Gebhard, S. Bachmann. Facial expressionof emotions in borderline personality disorder and depression. J. Behav.Ther. Exp. Psychiatry, 36, pp. 183196, 2005.

[22] R. Rosenthal. Conducting judgment studies: Some methodological is-sues. The handbook of methods in nonverbal behavior research, pp.287365, 2005.

[23] E. Sariyanidi, H. Gunes, A. Cavallaro, Automatic analysis of facialaffect: A survey of registration, representation and recognition. IEEETransactions on PAMI, 37(6), pp.1113-1133, 2015.

[24] Y. Tong, W. Liao, Q. Ji Facial action unit recognition by exploiting theirdynamics and semantic relationships IEEE Transactions on PAMI, 29,pp. 16831699, 2007.

[25] E. Z. Tronick. Emotions and emotional communication in infants. AmPsychol, 44(2), pp 112-119,1989.

[26] M.F. Valstar, M. Pantic Fully automatic recognition of the temporalphases of facial actions IEEE Trans. Syst., Man, Cybernet., Part B(2012), pp. 2843

[27] Michel F. Valstar, Bjrn Schuller, Kirsty Smith, Florian Eyben, BihanJiang, Sanjay Bilakhia, Sebastian Schnieder, Roddy Cowie, and MajaPantic, ’AVEC 2013 - The Continuous Audio/Visual Emotion andDepression Recognition Challenge’, Proc. 3rd ACM AVEC challenge,pp. 3-10, 2013

[28] Michel Valstar, Jeff Girard, Timur Almaev, Gary McKeown, MarcMehu, Lijun Yin, Maja Pantic, Jeff Cohn, FERA 2015 - Second FacialExpression Recognition and Analysis Challenge, Proc. IEEE Int’l Conf.Face and Gesture Recognition, 2015

[29] Z. Wang, S. Wang, Q. Ji, Capturing complex spatio-temporal relationsamong facial muscles for facial expression recognition, in: Proc. IEEECVPR, pp. 34223429, 2013.

[30] X. Xiong, F. de la Torre. Supervised descent method and its applicationsto face alignment. Proc. IEEE CVPR, pp. 532-539, June 2013.

[31] Zeng, Z., Pantic, M., Roisman, G.I., and Huang, T.S. A Survey of AffectRecognition Methods: Audio, Visual, and Spontaneous Expressions,IEEE Transactions on PAMI, 31(1), pp. 39-58, 2012.