Judgment of perceived exertion by static and dynamic facial expression

Similar documents
Comparison of the relation with muscle activity (facial and leg) between Borg- and facial pictorial-rpe scale for perceived exertion

Exertional Observation in Adults Performing Intermittent Treadmill Walking and Running

Valence and Gender Effects on Emotion Recognition Following TBI. Cassie Brown Arizona State University

BUFFALO CONCUSSION BIKE TEST (BCBT) INSTRUCTION MANUAL

Trait Perceptions of Dynamic and Static Faces as a Function of Facial. Maturity and Facial Expression

CONCURRENT VALIDITY OF A PICTORIAL RATING OF PERCEIVED EXERTION SCALE FOR BENCH STEPPING EXERCISE. Maressa Priscila Krause

Validation of the Children s OMNI RPE Scale for Stepping Exercise

Validity of a pictorial perceived exertion scale for effort estimation and effort production during stepping exercise in adolescent children

RELIABILITY OF EFFORT PRODUCTION USING THE CHILDREN S CALER AND BABE PERCEIVED EXERTION SCALES

Pupil Dilation as an Indicator of Cognitive Workload in Human-Computer Interaction

Teaching Dating and Relationship Skills to Teenagers with High Functioning Autism

Description and explanation of the major themes of The Curious Incident of the Dog in the other people, dealing with new environments, and making

Understanding Facial Expressions and Microexpressions

Notebook computer use with different monitor tilt angle: effects on posture, muscle activity and discomfort of neck pain users

RISK COMMUNICATION FLASH CARDS. Quiz your knowledge and learn the basics.

(SAT). d) inhibiting automatized responses.

This paper is in press (Psychological Science) Mona Lisa s Smile Perception or Deception?

Running head: FACIAL EXPRESSION AND SKIN COLOR ON APPROACHABILITY 1. Influence of facial expression and skin color on approachability judgment

RUNNING HEAD: RAPID EVALUATION OF BIG FIVE 1

PROPEL: PRomoting Optimal Physical Exercise for Life* Submaximal Graded Exercise Assessment Guidelines

Effects of a home-based aerobic exercise program on glycosylated. hemoglobin and the peak oxygen uptake in children and. adolescents With Type 1 DM

Effect of Psycho-physical Stress on the Preference of Non- Athletic Youths for Intermittent or Continuous Bench- Stepping

Psychology Perception

Fukuoka University of Education

Talking about an issue may not be the best way

Concurrent Validity of Borg s Rating of Perceived Exertion in African-American Young Adults, Employing Heart Rate as the Standard

Person Perception. Forming Impressions of Others. Mar 5, 2012, Banu Cingöz Ulu

ACGIH TLV for Hand Activity Level (HAL)

ID# Exam 3 PS 217, Spring 2009 (You must use your official student ID)

AFFECTS OF MUSIC ON EXERCISE. Name Institution

Grace Iarocci Ph.D., R. Psych., Associate Professor of Psychology Simon Fraser University

Recognition of Faces of Different Species: A Developmental Study Between 5 and 8 Years of Age

The effects of music tempo on cycling performance. R. Appell, K. Carnes, S. Haase, C. Haia, E. Smith, K. Smith, and J. Walsh

BURNOUT The Overtraining Syndrome in Swimming

PERCEPTION and COMMUNICATION. Week 2

Nonverbal Communication6

Facial Dynamics as Indicators of Trustworthiness and Cooperative Behavior

Emotional Development

The Effects of Upper -Body and Lower -Body Fatigue on Standing Balance

Who Needs Cheeks? Eyes and Mouths are Enough for Emotion Identification. and. Evidence for a Face Superiority Effect. Nila K Leigh

Prince Willem Alexander and Princess Maxima. Analyzed according to the Twelve Goodfield Personality Types. by Prof. Barry A. Goodfield, Ph.D.

Melinda R. Bolgar. B.A., Ithaca College, M.E.S.S. & M.S., University of Florida, 2004 & Submitted to the Graduate Faculty of

A Method for Non-experts in Assessing Exposure to Risk Factors for Work-related Musculoskeletal Disorders ERIN

California Subject Examinations for Teachers

VALIDATION OF ADULT OMNI PERCEIVED EXERTION SCALES FOR ELLIPTICAL ERGOMETRY

The Relation Between Reactive Strength Index and Running Economy in Long-Distance Runners Nicholas Gallina Dr. David Diggin

Facial Expression Biometrics Using Tracker Displacement Features

CHAPTER THIRTEEN Managing Communication

Ergonomic Test of the Kinesis Contoured Keyboard

Running head: INDIVIDUAL DIFFERENCES FOR EMOTION EXPRESSIONS 1. Individual Differences Are More Important

Olfactory Cues Modulate Facial Attractiveness. Demattè, Österbauer, & Spence 2007

An Ambiguous-Race Illusion in Children s Face Memory Kristin Shutts and Katherine D. Kinzler

Chapter 3. Perception and Memory. Factors That Affect Teacher s Guidance

Running Head: PERCEIVED REALISM OF DYNAMIC FACIAL EXPRESSIONS. Perceived Realism of Dynamic Facial Expressions of Emotion

MRI Ergometer Pedal. Medication free, load adjusted patient activation during MRI scans. Unique ergometer for MRI.

What is Emotion? Emotion is a 4 part process consisting of: physiological arousal cognitive interpretation, subjective feelings behavioral expression.

THE USE OF A 10-POINT EFFORT PERCEPTION SCALE IN ADULTS: A PRELIMINARY STUDY

Empirical Knowledge: based on observations. Answer questions why, whom, how, and when.

Accepted author version posted online: 11 Apr 2015.

Published in: Pediatric Exercise Science. Document Version Publisher's PDF, also known as Version of record

STANDARDISATION OF PHYSIOLOGY NOMENCLATURE Rebecca Tanner, Pitre Bourdon

Psychology Scientific Inquiry Domain Research Methods, Measurement, and Statistics

Mapping the Runner s Mind: A New Methodology for Real-Time Tracking of Cognitions

Exploring essential skills of CCTV operators: the role of sensitivity to nonverbal cues

Effects on an Educational Intervention on Ratings of Criminality, Violence, and Trustworthiness based on Facial Expressions By Lauren N.

UNIVERSITY OF WISCONSIN-LA CROSSE. Graduate Studies COMPARISON OF THE RATING OF PERCEIVED EXERTION (RPE) AND

Running head: EMPATHY, MIRROR NEURONS, and HEART RATE

FACIAL EXPRESSION RECOGNITION FROM IMAGE SEQUENCES USING SELF-ORGANIZING MAPS

The Vine Assessment System by LifeCubby

Contextual determinants of pain judgments

Tilburg University. Published in: Proceedings of the 13th International conference of the international Speech Communication Association (Interspeech)

CONSUMERS PREFERENCE ON SCOOTER DESIGN WITH GENDER- NEUTRAL STYLE

Virtual Reality-based Personalized Learning Environment for Repetitive Labor-intensive Construction Tasks

University of Groningen. Imitation of emotion Velde, Sytske Willemien van der

Clever Hans the horse could do simple math and spell out the answers to simple questions. He wasn t always correct, but he was most of the time.

From Confounders to Suspected Risk: The Role of Psychosocial Factors Michael Feuerstein, Ph.D., MPH

The Effects of Voice Pitch on Perceptions of Attractiveness: Do You Sound Hot or Not?

Article. Reference. The Perception and Mimicry of Facial Movements Predict Judgments of Smile Authenticity. KORB, Sebastian, et al.

Public Speaking Chapter 1. Speaking in Public

Stability of internal response and external load during 4-a-side football game in an indoor environment

SHORT REPORT Facial features influence the categorization of female sexual orientation

Perceptual and Motor Skills, 2010, 111, 3, Perceptual and Motor Skills 2010 KAZUO MORI HIDEKO MORI

Products for life and performance.

Self-Assessment: Critical Skills for Inclusion Practitioners Developed by Kathy Obear, Ed. D., 2014

A study of the effect of auditory prime type on emotional facial expression recognition

THE EFFECT OF CADENCE ON TIME TRIAL PERFORMANCE IN NOVICE FEMALE CYCLISTS. Patricia Graham. A Thesis Submitted to the Faculty of

In the Heat of the Moment

8/17/2012. Self-Concept Video. Cultural Differences in Defining the Self. Chapter 5. The Self: Understanding Ourselves in a Social Context

INTRODUCTION TO RELATIONSHIP DEVELOPMENT INTERVENTION (RDI )

Effects of Capital Collar Enhanced on Head-Cervical Movements in Comparison with Miami J Advanced and Aspen Vista TX Collars

Development of the Web Users Self Efficacy scale (WUSE)

Fundamentals of Cognitive Psychology, 3e by Ronald T. Kellogg Chapter 2. Multiple Choice

Gestalt Principles of Grouping

Computer based cognitive rehab solution

ASSOCIATIONS OF ACOUSTIC ENVIRONMENT WITH PHYSIOLOGICAL AND PSYCHOLOGICAL STATES AMONG HOSPITALIZED PATIENTS IN GENERAL HOSPITAL WARDS

Brain and Cognition, 48(2-3), (2002) Evaluation of nonverbal emotion in face and voice: some preliminary findings on a new battery of tests

Channel Dominance in Decoding Affective Nonverbal Behavior Displayed by Avatars 1. What You See is What You Get:

(Visual) Attention. October 3, PSY Visual Attention 1

Perceptual Disorders. Agnosias

When facial attractiveness is only skin deep

Transcription:

Judgment of perceived exertion by static and dynamic facial expression Ding Hau Huang a, Wen Ko Chiou a, Bi Hui Chen b a Graduate Institute of Business and Management, College of Management, Chung Gung University, Taiwan; b Department of Business Administration, Chihlee Institute of Technology, Taiwan Unlike facial expressions of emotion, the effect of static and dynamic facial information on the perception of physical exertion have not been discussed in much depth. The purpose of this study is to examine whether the observer could perceive exertion by others based only on static or dynamic facial expression. And to examine which conditions (static and dynamic exertion face) could be better perceived by observers. Fifteen male and 15 female observers were recruited to observe both static and dynamic exertional faces to estimate the perceived exertion by facial pictorial RPE scale. The results of this study suggested that observers could perceive exertion by others only based on static or dynamic facial expression. The dynamic characteristic of our facial display is an important factor in the perception of the low intensity of exertional expressions. The static exertional face of a male is better than the dynamic exertional faces for perceiving incense facial expression. Practitioner Summary: This study demonstrated that observers could use the facial pictorial RPE scale to assess both static and dynamic facial expression of effort. The effect of static and dynamic display could be used by ergonomist for observational assessment of forceful exertion. Keywords: Facial expression of effort, perceived exertion, dynamic and static stimuli 1. Introduction The rating of perceived exertion (RPE) scale has been widely used for workers to measure their own exertion. This method is commonly used for subjectively evaluating work tasks and determining acceptable loads in order to understand the task demands and prevents injuries. Besides exertion assessment by workers themselves, level of forceful exertions can also be estimated through observations by others. Observational methods offer an advantage in that they do not interfere with the task, thereby minimizing disruption. As Robertson et al. (2006) suggested that exertional observation is unobtrusive, inexpensive, and has a low participant burden making it suited for use in physical activity surveys, exercise prescription, and fitness-tracking involving children and adolescents. Some studies have demonstrated that observers could use RPE scales to assess someone else s exertion. For examples, Ljunggren (1986) examined concurrent validity of a observational method to estimate Borg RPE of subjects performing randomly presented cycle ergometer power outputs. Drury et al. (2006) observed subjects who performed different postures for weight holding task to estimate Borg RPE. Robertson et al. (2006) showed the direct kinematic observation procedure to estimate OMNI Scale RPE for female and male children performing treadmill exercise. These studies have shown that observer-estimated RPE was strongly correlated with the subject s RPE. Usually exertion observation needs some visual cues for measuring forceful exertion in terms of detecting differences in exertion intensity. Facial expression is believed to be an important observation key in the observational assessment of forceful exertion that could provide a perceptual estimate of exertion intensity (Huang and Chiou, 2013; Huang et al., 2014; de Morree and Marcora, 2010). The face of effort is a type of facial expression that reflects how hard people are working during physical tasks and exercise. Previous studies have been shown that facial muscle activity correlated with RPE during weightlifting and cycling (de Morree and Marcora, 2010; Huang et al., 2014). Facial expression not only can convey emotion but also express physical effort; grimacing and squinting are examples of non-verbal behavior used to judge effort. Most research on the facial expressions of emotion has been conducted using static faces as stimuli. However, the importance of dynamic properties of facial expression has recently been emphasized in some psychological literature. Some studies have shown a particular sensitivity for dynamic facial movements such 1

as, recognizing emotional expressions (Kamachi et al., 2001; Ambadar et al., 2005), in the perceived intensity of basic emotional expressions (Biele and Grabowska, 2006), recognizing mental states and even in the perceived intensity of of pain expression (Simon et al., 2008). Nevertheless, some studies also indicated that there was no difference between static and dynamic facial perception. Recio et al. (2011) showed that no significant differences in error rates of recognizing were observed between static and dynamic stimuli within neutral and anger facial expression. Rhodes et al. (2011) revealed that facial attractiveness ratings from video-clips and static Images tell the same story. Even though the videos contained a lot more information than the photographs, which could potentially affect perceptions of attractiveness. However, either static or dynamic faces on facial expression of effort have not been discussed in much depth, and an effect of static and dynamic facial information on perceived of physical exertion remains to be determined. Accordingly, the purpose of this study is to examine weather observer could perceive exertion by others based only on static or dynamic facial expression. And to examine which conditions (static and dynamic exertion face) could be better perceived by observers. 2. Method 2.1 Experimental design A cross-sectional, perceptual estimation design was used to assess exertional perceptions during a loadincremented cycle ergometer protocol that terminated at peak task intensity. Each participant performed continuous incremental workload cycling exercise on an electromagnetically braked cycle ergometer (Corival 906900, Lode BV, Groningen, The Netherlands). The exercise included a 5-minute warm up at 0 watts (W), the initial power output was 50 W for women and 75 W for men. Power outputs were incremented in continuous 3-min test stages by 25 W and 50 W, respectively, for women and men. Participants were instructed to maintain a cycling cadence between 60 70 rpm throughout the exercise test. When a participant experienced volitional exhaustion or an inability to sustain a cadence greater than 55 rpm for a period of 5 seconds, the exercise test was terminated. A Sony digital video camcorder (HDV 1080i- HV20) was used, at a fixed distance of 60 cm from the subject s face, at eye level. The final videos were edited to 10 seconds and muted, using imovie 4.0 from each stage. During the final 30 seconds of each stage (3 minutes) of the cycle ergometer exercise, the participants were asked to report their feelings of exertion according to the facial pictorial RPE scale (Huang and Chiou, 2013). An undifferentiated rating was estimated for the overall body. The facial pictorial RPE scale has a category rating format that contains both pictorial and verbal descriptors positioned along a comparatively narrow numerical response range, namely 0 10. The facial pictorial RPE scale was chosen because of the visual information on the pictorial faces was the primary cues emphasizing the scale instructions and this scale has been evidence supported by the use of this facial pictorial RPE scale by adult females and males for estimating RPE. HR was monitored continuously using a wireless chest strap telemetry system (Polar WearLink System and Polar FT4 HR monitor, Polar Electro Oy, Kempele, Finland) during the task. Throughout the test the corrugator supercilii (CS) muscle activities were continuously recorded using EMG during exercise. For details about the recording method, please see our previous publication on facial EMG (Huang et al., 2014). Five of the 22 male and 4 of the 18 females subjects were selected to analyze their video clips because of they were able to complete at least five stages of exercise. Dynamic stimulus were created by editing the videos to 6-s clips at the final 30 s of each stage and saving them in Quicktime format at 30 frames per second. Static stimuli were created by taking from each video a frame in which the frame was expressly exposed and the face maintained an actual state of exertion, looking toward the camera, and without purposely offering extra emotional expression. All faces were displayed for rating on an imac 27 inch PC and viewed from 57 cm. Written informed consent was obtained from all participants. 2.2 Observation procedure Thirty observers were recruited including 15 male and 15 female. Each observer observed both static and dynamic exertional faces to estimate perceived exertion by facial pictorial RPE scale. Randomize the display order of each subject s faces into 5 groups. Observers randomly selected one group to observe. The first 2

face displayed was selected from the warm up stage where no effort was made in order to show the baseline condition of each subject. After a baseline face was observed, the faces from each stimulation stage were assessed. The printed facial pictorial RPE scale was for observers to estimate the RPE when the faces were shown. 2.3 Data analysis Subject s self-report RPE was categorized into three levels. Subject s RPE from 0 to 3 was low RPE level, 4-6 was medium RPE level, and 7-10 was high RPE level. Descriptive analysis of physiology and RPE data were shown as means ± sd. Statistical analyses was performed using a two-way ANOVA to examine the difference between RPE levels and using two-tailed student s t-test to examine the difference of estimated error between static and dynamics display conditions. The correlations between RPE estimated by the observer and RPE estimated by the subject were examined using a Pearson product moment correlation coefficient. Statistic method Significance was set at a P value of 0.05 (two-sided) for all analyses. All analyses were conducted using the Statistical Package for Social Sciences Version 20 (SPSS Inc., Chicago, IL) 3. Result Table 1 presents the average and standard deviation of physiology and RPE data of three RPE levels. Physiology data, HR and facial muscle activity of subjects for both gender were all significantly different between RPE levels except facial activity of female. Facial muscle activity of females did not change significantly through the task. All RPEs showed significant difference between RPE levels for both genders. Both static and dynamic observer RPEs were under estimated. Overall, the estimated error of observing female faces was significantly higher than male faces for both display conditions. Table 1. The average and standard deviation of physiology and RPE data of three RPE levels. Subject gender Male Female RPE level CS muscle Subject s Static Dynamic HR of subject activity RPE RPE RPE Low (Easy) 112.0±20.0 4.95%±2.1% 2.25±0.4 2.87±1.5 1.88±1.4 Medium (somewhat hard) 130.1±25.5 12.78%±5.3% 5.11±0.7 4.12±1.8 3.00±1.8 High (Hard) 164.5±16.0 29.47%±16.9% 8.25±1.0 7.41±1.8 6.90±2.1 Low (Easy) 110.5±12.9 11.80%±10.0% 2.75±0.4 2.16±1.8 1.91±1.1 Medium (somewhat hard) 142.1±15.8 18.06%±16.7% 4.80±0.8 3.07±1.9 2.88±1.7 High (Hard) 170.2±9.0 18.70%±13.3% 7.40±0.5 5.55±2.4 6.07±1.9 Table 2 shows that for both genders, static and dynamic RPE estimated by observers were all significantly correlated with subject s self report RPE. Otherwise, all the RPEs were significantly correlated with HR. Comparisons between two observers RPEs, dynamic RPE show higher correlation between subject s RPE and HR than static RPE. 3

Table 2. Correlation between subject s RPEs, HR and observer s RPE Male HR Subject s RPE Subject s RPE 0.785** Static RPE 0.562** 0.718** Dynamic RPE 0.604** 0.752** Female Subject s RPE 0.835** Static RPE 0.649** 0.529** Dynamic RPE 0.654** 0.661** In table 3, there was no gender difference between observers to estimate subject s RPE. However, male observers have the tendency of a smaller error of judgment than males for both display conditions but no significant difference. Although there is no gender effect from observers, the gender difference between subjects was matter. Overall, the estimated error by observing female subjects was significantly higher than observing male subjects for both display conditions. Table 3. Sex difference in estimated error of observers Subject Gender All Male Female Display conditions Static Dynamic Static Dynamic Static Dynamic Observer gender Male 1.71±1.36 Female 1.84±1.45 Male 1.84±1.37 Female 1.96±1.43 Male 1.44±1.27 Female 1.50±1.29 Male 1.86±1.38 Female 1.97±1.48 Male 2.00±1.39 Female 2.21±1.53 Male 1.82±1.36 Female 1.94±1.37 Mean±SD t P -1.60 0.11-1.42 0.16-0.59 0.55-0.97 0.33-1.69 0.09-1.04 0.30 Figure 1 showed the estimated error (observer estimate subject s self-report) associated with observers estimation of exertion expression through static and dynamic faces for males and females. A two-way analysis of variance (ANOVA), with factors perceived exertion levels of subjects (low, medium and high) and display type (dynamic, static), revealed main effects of perceived exertion level, display type, and an interaction between these factors for both genders. For males, the subject s RPE are at low level, estimated error of dynamic face was significantly lower than static face. At medium and high level of estimated error of dynamic face was significantly higher than static face. For females, the subject s RPE at low and high level, estimated error of dynamic face was significantly lower than static face. There was no significant difference between static and dynamic condition when female are at medium level. 4

Figure 1. Estimated error of observers from three RPE levels of subjects 4. Discussion 4.1 Observation exertion by facial expression This study hypothesized that both RPE of face display condition estimated by the observer and subject would be positively correlated. The findings supported this hypothesis, the observer estimated RPE through facial pictorial RPE scale which was significantly correlated with the subject s self-rated RPE who used the same scale. This result indicated that only observing static or dynamic facial expression for perceiving others exertion could be possible and the correlation coefficient of dynamic RPE would be slightly higher than static RPE. The result of significant correlation was consistent with previous studies by using different RPE scales. Observer-estimated RPE was positively and linearly related to subject s self-rated RPE through OMNI RPE scale over the submaximal cycle ergometer exercise stages (r=0.87 0.92) (Robertson et al., 2006). Ljunggren (1986) showed that to estimate RPE of subjects performing cycle ergometer power outputs through Borg scale. The observer-estimated RPE and subject s self-rated RPE were also highly correlated (r=0.99). Drury et al. (2006) also showed that observers-estimated RPE through Borg scale correlated highly 5

with the self-rated RPE of the subjects performing the static holding tasks. However, the correlation between both observer RPE and subject s RPE of this present study was significantly correlated (r=0.56 0.65) but lower than above studies. The result may be explained by the type of observation, in previous studies observers estimated exertions by dynamic observation cues with direct observations or video clips such as body language, facial expression and postures. Although dynamic exertion faces was also observed in this present study, only facial cues remained limited. The detail difference between static and dynamic is discussed in next section. 4.2 The difference between static and dynamic display conditions for observing The results of this study showed that static face of exertion seems to have less estimated error than dynamic. However, from the view of the three subject s RPE levels, when both gender subjects feel easy (subject s RPE from 0-3), observers estimated error through their dynamic faces would be significantly less than static one for both male and female observers. One explanation for this is that dynamic facial expression offering more information to estimate the slight changes in facial activity when subjects are perceiving low-level exertion. The result is in line with previous studies of emotion recognition. Bould and Morris (2008) demonstrated that dynamic facial expressions could be a useful cue in the recognition process, especially when the expressions are subtle. Ambadar et al. (2005) also suggested that dynamic presentation might be particularly beneficial for the recognition of emotional expressions at low intensity. When male subjects reported somewhat hard (subject s RPE from 4-6) or hard (subject s RPE from 7-10) they were facially expressive, the observer-estimated error through these dynamic faces would be significantly higher than static one. The more likely explanation rets in too much confusing information for exertion estimation through dynamics faces. As Bould and Morris (2008) suggested that when an emotion is intense, a static face is able to provide sufficient information to recognize an emotional state of someone else from a still photograph. It therefore seems that static faces are already strong carriers of emotional signals by corresponding to the shared stereotypes, leaving little scope for improvement through the provision of dynamic information. Other studies also suggested that the dynamic facial expression was not always better than static, for instant, facial attractiveness rating (Rhodes et al., 2011), emotion recognition of anger (Recio et al., 2011) and mental status recognition (Back et al., 2009). Therefore, the effect of dynamic face depends on different situations. On the other hand, when female subjects reported hard during task, the observer-estimated error through dynamic faces would be much less than observing static faces. This result is in contrast with male subjects, this may be explained that females didn t change facial activity obviously even when they perceived hard. Therefore, whether male and female subjects whom were not facially expressive, the dynamic facial expression could be a better cue in the rating of perceived exertion. 4.3 Gender difference of perceiving exertion The present study showed that there was no gender difference between observers to estimate subject s RPE through both display conditions. Male observers through both display conditions have the tendency of a smaller error of judgment than females but no significant difference. The result wasn t in line with previous emotional studies. Previous studies revealed that females have an advantage in the decoding of non-verbal emotional cues and have more sensitivity perceiving facial expression of emotion (Hall et al., 2000; McClure, 2000; Montagne et al., 2005). One explanation is that the difference of neural mechanism of empathy between emotional processing and exertion perceiving. Future research could examine the difference between neural mechanism of emotion and exertion. 5. Conclusion The results of this study suggested that observers could perceive exertion by others only based on static or dynamic facial expression. Type of face display conditions influenced on rating of perceiving others RPE. Its effect, however, depends on the subject s sex and level of subject s RPE. The dynamic characteristic of facial display is an important factor in the perception of the low intensity of exertional expressions. The static 6

exertional face of males had less estimated error than dynamic exertional faces for perceiving incense facial expression. References Ambadar, Z., Schooler, J. W., and Cohn, J. F. 2005. "Deciphering the enigmatic face the importance of facial dynamics in interpreting subtle facial expressions". Psychological Science 16(5): 403-410. Back, E., Jordan, T. R., and Thomas, S. M. 2009. "The recognition of mental states from dynamic and static facial expressions". Visual Cognition 17(8): 1271-1286. Biele, C., and Grabowska, A. 2006. "Sex differences in perception of emotion intensity in dynamic and static facial expressions". Experimental Brain Research 171(1): 1-6. Bould, E., and Morris, N. 2008. "Role of motion signals in recognizing subtle facial expressions of emotion". British Journal of Psychology 99(2): 167-189. de Morree, H. M., and Marcora, S. M. 2010. "The face of effort: frowning muscle activity reflects effort during a physical task". Biological Psychology 85(3): 377-382. Drury, C. G., Atiles, M., Chaitanya, M., Lin, J.-F., Marin, C., Nasarwanji, M., et al. 2006. "Vicarious perception of postural discomfort and exertion". Ergonomics 49(14): 1470-1485. Hall, J. A., Carter, J. D., and Horgan, T. G. 2000. "Gender differences in nonverbal communication of emotion". Gender and emotion: Social Psychological Perspectives: 97-117. Huang, D. H., and Chiou, W. K. 2013. "Validation of a facial pictorial rating of perceived exertion scale for evaluating physical tasks". Journal of Industrial and Production Engineering 30(2): 125-131. Huang, D. H., Chou, S. W., Chen, Y. L., and Chiou, W. K. 2014. "Frowning and jaw clenching muscle activity reflects the perception of effort during incremental workload cycling". Journal of Sports Science & Medicine 13(4): 921. Kamachi, M., Bruce, V., Mukaida, S., Gyoba, J., Yoshikawa, S., and AkamatsuÙ, S. 2001. "Dynamic properties influence the perception of facial expressions". Perception 30: 875-887. Ljunggren, G. 1986. "Observer ratings of perceived exertion in relation to self ratings and heart rate". Applied Ergonomics 17(2): 117-125. McClure, E. B. 2000. "A meta-analytic review of sex differences in facial expression processing and their development in infants, children, and adolescents". Psychological Bulletin 126(3): 424. Montagne, B., Kessels, R. P., Frigerio, E., de Haan, E. H., and Perrett, D. I. 2005. "Sex differences in the perception of affective facial expressions: Do men really lack emotional sensitivity?" Cognitive Processing 6(2): 136-141. Recio, G., Sommer, W., and Schacht, A. 2011. "Electrophysiological correlates of perceiving and evaluating static and dynamic facial emotional expressions". Brain Research 1376: 66-75. Rhodes, G., Lie, H. C., Thevaraja, N., Taylor, L., Iredell, N., Curran, C., et al. 2011. "Facial attractiveness ratings from video-clips and static images tell the same story". PloS One 6(11): e26653. Robertson, R. J., Goss, F. L., Aaron, D. J., Tessmer, K. A., Gairola, A., Ghigiarelli, J. J., et al. 2006. "Observation of perceived exertion in children using the OMNI pictorial scale". Medicine and Science in Sports And Exercise 38(1): 158. Simon, D., Craig, K. D., Gosselin, F., Belin, P., and Rainville, P. 2008. "Recognition and discrimination of prototypical dynamic expressions of pain and emotions". PAIN 135(1): 55-64. 7