Emote to Win: Affective Interactions with a Computer Game Agent

Similar documents
Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry.

Emotion based E-learning System using Physiological Signals. Dr. Jerritta S, Dr. Arun S School of Engineering, Vels University, Chennai

Emotion Detection Using Physiological Signals. M.A.Sc. Thesis Proposal Haiyan Xu Supervisor: Prof. K.N. Plataniotis

Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners

Affective Game Engines: Motivation & Requirements

Research Proposal on Emotion Recognition

On Shape And the Computability of Emotions X. Lu, et al.

Emotions and Motivation

Empathy for Max. (Preliminary Project Report)

Estimating Intent for Human-Robot Interaction

Motivation represents the reasons for people's actions, desires, and needs. Typically, this unit is described as a goal

PHYSIOLOGICAL RESEARCH

Towards an EEG-based Emotion Recognizer for Humanoid Robots

draft Big Five 03/13/ HFM

Mental State Recognition by using Brain Waves

Thought Technology Ltd.

Blue Eyes Technology

Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification

Recognising Emotions from Keyboard Stroke Pattern

Affective Dialogue Communication System with Emotional Memories for Humanoid Robots

Smart Sensor Based Human Emotion Recognition

Edge Based Grid Super-Imposition for Crowd Emotion Recognition

Emotion Affective Color Transfer Using Feature Based Facial Expression Recognition

A Cooperative Multiagent Architecture for Turkish Sign Tutors

Introduction to affect computing and its applications

Augmented Cognition to enhance human sensory awareness, cognitive functioning and psychic functioning: a research proposal in two phases

Real Time Sign Language Processing System

Affect in Virtual Agents (and Robots) Professor Beste Filiz Yuksel University of San Francisco CS 686/486

Biofeedback is a new tool and you don t need expensive equipment to use it.

A Cooperative Multiagent Architecture for Turkish Sign Tutors

Recognizing, Modeling, and Responding to Users Affective States

MODULE 41: THEORIES AND PHYSIOLOGY OF EMOTION

Artificial Emotions to Assist Social Coordination in HRI

Detecting emotion from EEG signals using the emotive epoc device

Bayesian Networks for Modeling Emotional State and Personality: Progress Report

INTER-RATER RELIABILITY OF ACTUAL TAGGED EMOTION CATEGORIES VALIDATION USING COHEN S KAPPA COEFFICIENT

Contrastive Analysis on Emotional Cognition of Skeuomorphic and Flat Icon

Towards Human Affect Modeling: A Comparative Analysis of Discrete Affect and Valence-Arousal Labeling

Modeling the Use of Space for Pointing in American Sign Language Animation

Emotion Recognition using a Cauchy Naive Bayes Classifier

BIOSIGNALS ANALYSIS AND ITS APPLICATION IN A PERFORMANCE SETTING Towards the development of an Emotional-Imaging Generator

COMPUTER PLAY IN EDUCATIONAL THERAPY FOR CHILDREN WITH STUTTERING PROBLEM: HARDWARE SETUP AND INTERVENTION

What is Emotion? Emotion is a 4 part process consisting of: physiological arousal cognitive interpretation, subjective feelings behavioral expression.

MEASURING EMOTION: A NEW EVALUATION TOOL FOR VERY YOUNG CHILDREN

Facial expression recognition with spatiotemporal local descriptors

Emotions. These aspects are generally stronger in emotional responses than with moods. The duration of emotions tend to be shorter than moods.

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

MSAS: An M-mental health care System for Automatic Stress detection

Social Signal Interpretation (SSI):

Emotionally Augmented Storytelling Agent

PERFORMANCE EVALUATION OF VARIOUS EMOTION CLASSIFICATION APPROACHES FROM PHYSIOLOGICAL SIGNALS

Effect of Visual Feedback Caused by Changing Mental States of the Avatar based on the Operator s Mental States using Physiological Indices

Pattern Classification

Emotion Analysis Using Emotion Recognition Module Evolved by Genetic Programming

The Ordinal Nature of Emotions. Georgios N. Yannakakis, Roddy Cowie and Carlos Busso

Audiovisual to Sign Language Translator

NeuroSky s esense Meters and Detection of Mental State

MUSIC is important for people because it pleasure

General Psych Thinking & Feeling

Myers Psychology for AP*

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 5, JULY

The Schema is Mightier Than the Sword Using Player Cognition to Predict Gaming Behavior

Skin color detection for face localization in humanmachine

COGNITIVE CONTROL IN MEDIA MULTITASKERS

Facial Expression Biometrics Using Tracker Displacement Features

EIQ16 questionnaire. Joe Smith. Emotional Intelligence Report. Report. myskillsprofile.com around the globe

Facial Expression Recognition Using Principal Component Analysis

Introduction to Psychology. Lecture no: 27 EMOTIONS

Development of 2-Channel Eeg Device And Analysis Of Brain Wave For Depressed Persons

Anxiety Detection during Human-Robot Interaction *

Inventions on expressing emotions In Graphical User Interface

Motivation and Emotion. Unit 2: Biopsychology

EECS 433 Statistical Pattern Recognition

A USER INDEPENDENT, BIOSIGNAL BASED, EMOTION RECOGNITION METHOD

HTS Report EPS. Emotional Processing Scale. Technical Report. Anna Patient ID Date 29/09/2016. Hogrefe Ltd, Oxford

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

Selection of Emotionally Salient Audio-Visual Features for Modeling Human Evaluations of Synthetic Character Emotion Displays

Emotional Design. D. Norman (Emotional Design, 2004) Model with three levels. Visceral (lowest level) Behavioral (middle level) Reflective (top level)

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face

Bio-Feedback Based Simulator for Mission Critical Training

A Pilot Study on Emotion Recognition System Using Electroencephalography (EEG) Signals

Observer Annotation of Affective Display and Evaluation of Expressivity: Face vs. Face-and-Body

Audio-visual Classification and Fusion of Spontaneous Affective Data in Likelihood Space

Motivation. The Emotion is the way of the representation to express their inner face. - Bertand Russell, Philosopher, of 24

This is the accepted version of this article. To be published as : This is the author version published as:

Temporal Context and the Recognition of Emotion from Facial Expression

SPEECH EMOTION RECOGNITION: ARE WE THERE YET?

N RISCE 2K18 ISSN International Journal of Advance Research and Innovation

Elements of Communication

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

Expression of basic emotion on playing the snare drum

VITHEA: On-line word naming therapy in Portuguese for aphasic patients exploiting automatic speech recognition

Mindfulness & Relationships

FILTWAM. A Framework for Online Affective Computing in Serious Games. Kiavash Bahreini Wim Westera Rob Nadolski

EMOTIONAL INTELLIGENCE QUESTIONNAIRE

Running head: HEARING-AIDS INDUCE PLASTICITY IN THE AUDITORY SYSTEM 1

Real-time Recognition of the Affective User State with Physiological Signals

Motion Control for Social Behaviours

Transcription:

Emote to Win: Affective Interactions with a Computer Game Agent Jonghwa Kim, Nikolaus Bee, Johannes Wagner and Elisabeth André Multimedia Concepts and Application, Faculty for Applied Computer Science University of Augsburg, Eichleitnerstr. 3, 8659 Augsburg, Germany {jonghwa.kim, andre}@informatik.uni-augsburg.de Abstract:Inthispaper,weintroduceagameinterfacethatisbasedonaffectiveinteractions between a player and a computer pet. As opposed to many traditional computer games, users are not expected to manipulate control devices in a skillful manner to win thegame.insteadthebasicideaistoelicitcertainreactionsofthepetviaappropriate emotive user behaviors. For improved accuracy of emotion recognition, we employ a combined analysis of signals from two emotive expression channels: affective speech and physiological reactions. Introduction With advances in the area of sensor technology and signal processing, a new generation of computer games based on biofeedback and affective computing is emerging. In order to increase a player s level of immersion and engagement(see[fr3]), a number of affective games encourage a user to express his or her emotive states and dynamically adapt actions andeventstothem. InGirland, younggirlsmayuseanemotionwheeltoinputtheir emotive states, see http://www.girland.com/. In SenToy, the user interacts with a tangible dolltocommunicateoneofsixemotionsthroughgestures(see[pcp + 3]).Gamesrelying onbiofeedbackaimathelpingtheusertogaincontroloverhisorherbodilyreactions(e.g. see[lsb + 3]).Forinstance,inthegame TheWildDivine theplayerneedstoachieve a state of relaxation or concentration to manipulate devices of the game environment, see http://www.wilddivine.com/. Another example of an affective interface is the computer gamefinfinwheretheusermayinfluencetheemotionalstateofahalfbird,halfdolphin creature via talking and waving. Inthispaperweproposeanaffectiveuserinterfacetoacomputergamebasedonanew paradigmcalled EmotetoWin. Thebasicideaistoinfluencethebehaviorofavirtual pet- Tiffany, the snail- by appropriate emotional reactions of the user. While interacting with the snail, the user is monitored by means of bio sensors, measuring skin conductivity, heart rate, respiration and muscle activity. In addition, we collect information on the user s emotional state by analyzing his or her speech input. Tiffany, the snail, is not supposed WewouldliketothankthestudentsoftheMultimediaStudentProject SpielmitEmotionen fortheirwork on the implementation of Tiffany. 59

tounderstandtheuser sspeechinput. However,similartoarealpet,itissensitivetothe user s emotive state and tries to respond to it accordingly. As in The Wild Divine, there areseveraloptionstoplaythegame. Forinstance,theusermayaimatgettingTiffanyto performcertainactionswithinalimitedtime(tohaveakindofwinningconditionasin traditionalgames)orcommunicatewithitinanopen-endmodeinordertolearnhowto control his or her emotional response. The central components of our game environment are: modules for the recognition of emotions from biosignals and speech, a decision module for fusing the results from the single input channels and finally a module for determining and animating the behaviors of the virtual snail. For signal processing, Matlab was used. To enable real-time interactions, Tiffany s Java-based animation engine was linked with the emotion recognition system via Simulink/Matlab. Figure provides an overview of the system. USER CH. Speech 225 Hz/6bits CH 2. Biosignal (ECG, SC, RESP, EMG) Feature Extraction Feature Extraction Local Classifier (threshold-based) Local Classifier (threshold-based) Decision E (x, y) arousal Angry Happy valence Sad Calm Applications Figure : Conceptual Overview of the Affective Game Environment 2 The Underlying Emotion Model Two represent emotions, we follow a dimensional approach[la2] which characterizes emotions in terms of several continuous dimensions, such as arousal or valence. Arousal refers to the intensity of an emotional response. Valence determines whether an emotion is positive or negative and to what degree. A given emotional state is then characterized byapointinthistwodimensionalspace.emotiondimensionscanbeseenasasimplified representation of the essential properties of emotions. For instance, anger can be described by high arousal and negative valence. Apart from the ease of describing emotional states that cannot be distributed into clearcut fixed categories, the two dimensions valence and arousal are well suited for emotion recognition. As a first step, we aim at recognizing four distinct emotional states that may be associated with one of the four quadrants of our dimensional emotional model: angry (valence: negative, arousal: high), calm(valence: positive, arousal: low), sad(valence: negative, arousal: low) and happy(valence: positive, arousal: high). 3 Two-Channel Emotion Recognition Recently, there has been a signicant amount of work on the recognition of emotions(see [CDCT + ]foranoverview).mostapproachestoemotionrecognitionsofarconcentrate 6

onasinglemodalityanddonottakeadvantageofthefactthatanintegratedmultimodal analysis may help to resolve ambiguities and compensate for errors. An exception is the workbychenandcolleagues[cth + 98]whopresentanapproachtoemotionrecognition from speech and video, showing that higher recognition rates can be achieved by a bimodal analysis. The great challenge of our work is to recognize emotional signals from multiple channels while the player interacts with the application and to respond to them in realtime. There are several advantages of using biosensor feedback in addition to affective speech. First of all, we can continuously gather information on the user s emotive state through biosensors while the analysis of emotions from speech should only be triggered when the microphone receives speech signals from the user. Secondly, it is much harder for the user to deliberately manipulate biofeedfack than external channels of expression. 3. Data Collection To identify features from biosignals and speech whose values may be updated with reasonable computational cost, we collected a corpus containing physiological and speech data. In particular, we presented four subjects with a number of video sequences that are supposedtoevokeoneofthefouremotionsmentionedaboveandaskedthemforverbal feedback. While the subjects watched the video sequences, their physiological feedback was recorded using four biosensors. The ECG sensor(electrocardiogram) was used to measure the heart rate. The EMG sensor(electromyogram) determines the activity of muscles in the forehead to detect whether or not the user is frowning. The RESP sensor(respiration rate) captures abdominal breathing. The SC sensor(skin conductivity) measures sweat secretion, which was taken at the index and ring finger of the non-dominant hand. During the experiment, we also gathered the speech signals resulting from spontaneous verbal responses to the video sequences. 3.2 Feature Selection and Classification Therehasbeenagreatamountofworkinthedevelopmentofofflinerecognitionalgorithms based on discriminative or generative classifier models, such such as SVM(support vector machine), HMM (hidden markov model, and MLP (multilayer perception), see [CDCT + ]. Unfortunately,thesealgorithmscannotbedirectlyadaptedtothechallengingtaskofonlinerecognitionwheretheemotionalstateneedstoberecognizedassoonas possible(see[vp99] for discussion of these problems). As a first step, we decided to start from simple threshold-based classification methods. Secondly, we relied on a fixed set of pre-selected features based on an offline analysis of the collected data. From the speech signal, we extracted 23 features derived from pitch, harmonics and energy. From the biosignal, we extracted 25 features including derivations of mean energy from SC/EMG, standard deviation from SC/EMG, and four subband spectra from 6

4 3 2 5.5 5 4.5 2-36.5 36 3 2 3.8 3.6 3.4 3.2 3 2 - -.4 -.4 4 3 2 5 4.5 4 3.5 2-36 3 2 5.2 5 4.8 4.6 4.4 4.2 2-36 34 -.4 -.4 RESP/ECG. In Figure 2, examples of some features are shown with raw signals. Table shows how the recognition of emotions may profit from a combined analysis. For instance, skin conductance seems to be a good indicator for arousal. Determining the valenceofaemotionseemstobemuchtougher.however,thevoiceharmonicsmayhelp to distinguish, for example, positive emotions with high arousal from negative emotions with high arousal. 4 Affective Behavior of the Game Character Information on the user s emotive state is forwarded to Tiffany in terms of two coordinates that represent a point in the dimensional emotion model (x: valence, y: arousal). To respond to the user s emotional state, information on the user s emotive state is mapped onto appropriate facial and body movements of the snail(see Fig. 3). In addition, the snail isabletoexpressemotionsviasound.forinstance,theusermayscarethesnailawayby shouting at her angrily. As a consequence, Tiffany will hide behind a bush. In case Tiffany does not receives any utilizable input from the recognition module, for instance if the user doesnotshowanyinterpretablereactionoriftheemotionanalysisdoesnotleadtoany consistent results, certain idle time behaviors, such as eye blinking, will be elicited. Note thatthereisabi-directionalflowofemotionsbetweentheuserandthevirtualpet. The snail does not only respond to the user s emotive behaviors, but in turn evokes emotions intheuseraswell. Forexample,theusermaygetirritatedorboredifTiffanyrefusesto emerge from behind the bush. Happy 5 emg.4 emgmean ( s ec) s c rs p s cs lopemean ( s ec) 2 FreqMean ( s ec) rs pfreqmean ( s ec) 5.5 s c Angry emg.4 emgmean ( s ec) rs p s cs lopemean ( s ec) 2 FreqMean ( s ec) rs pfreqmean ( s ec) Sad emg.4 emgmean ( s ec) s c 36 rs p s cs lopemean ( s ec) 2 FreqMean ( s ec) rs pfreqmean ( s ec) emg 5.4 s c rs p Calm.4 emgmean ( s ec) s cs lopemean ( s ec) 2 FreqMean ( s ec) rs pfreqmean ( s ec) Figure 2: Biosignal Examples with Selected Features 62

ECG RESP SC EMG Pitch Harmonics Energy (heart rate) (Resp. Freq.) (Mean Value) (Mean Value) (F) (Fk-Cand.) (Mean Value) Happy + + + + ++/- - + Angry ++ + ++ + ++ + ++ Sad + +/- - - - -- -- Calm -- - -- - +/- +/- - Table : Emotional Cues in Speech and Biosignal Figure 3: The Emotion Character Tiffany 5 Discussion and Conclusion In this paper, we presented a first prototype of a gaming application for exploring new intuitive control strategies based on biophysiological feedback and affective speech. The game is based on the player s ability to express him- or herself emotionally. Our experience has shown that affective speech and biofeedback can be integrated in a computer game in a natural manner. As opposed to the user interfaces of many traditional computer games, the user did not have to learn complex control sequences for manipulating the game environment. Nevertheless, we identified a number of usability problems. Eventhoughthegamelogicwasquitesimple,thebehaviorofthevirtualsnailwasnot always obvious to the user. By their very nature, sensor data are heavily affected by noise and especially bio sensor systems are, among other things, very sensitive to motion artefacts. As a consequence, it was hard to acquire reliable information on the user s emotive state. To improve the recognition rate of our system, we are currently collecting a more comprehensive corpus of bio sensor and speech data. In addition, we are performing experiments with selected statistical clustering methods for offline analysis and investigate inhowfartheycanbeadaptedtoonlineanalysis. Wealsonoticedthatsomeusershadproblemswiththecontrolmechanismsofthegame since the mapping between their emotive input and Tiffany s resulting action was not alwayscleartothem. Forinstance,someoftheusersexpectedTiffanytoreflecttheirown emotions as opposed to responding to them. Others had problems to identify Tiffany s emotions from its bodily and facial behaviors. To remedy this, we intend to increase the transparency of the interface, e.g. by providing the user simple instructions on how to 63

influence Tiffany s behavior. Furthermore, we will exhibit Tiffany with more expressive animations and make more extensive use of sound. Weconcludethat EmotetoWin isaninterestingnewparadigmthatcanhelpusersto increase their awareness of their affective behavior and to deploy affect consciously to communicate(e.g., to control a game character). References [CDCT + ] Cowie,R.,Douglas-Cowie,E.,Tsapatsoulis,N.,Votsis,G.,Kollias,S.,Fellenz,W., und Taylor, J.: Emotion recognition in human-computer interaction. IEEE Signal Processing Magazine. 8():32 8. 2. [CTH + 98] Chen,L.,Tao,H.,Huang,T.,Miyasato,T.,undNakatsu,R.: Emotionrecognition from audiovisual information. In: Proceedings, IEEE Workshop on Multimedia Signal Processing. S. 83 88. Los Angeles, CA, USA. 998. [Fr3] [La2] Freeman, D.: Creating Emotion in Games: The Craft and Art of Emotioneering. New Riders. 23. Lang, P. J.: The emotion probe: Studies of motivation and attention. American Psychologist. 5(5):372 385. 22. [LSB + 3] Lyons,G.,Sharma,P.,Baker,M.,O Malley,S.,undShanahan,A.:Acomputergamebased emg biofeedback system for muscle rehabilitation. In: Engineering in Medicine and Biology Society, 23. Proceedings of the 25th Annual International Conference oftheieee.volume2.s.625 628.Sept.23. [PCP + 3] Paiva,A.,Chaves,R.,Piedade,M.,Bullock,A.,Andersson,G.,undHöök,K.:Sentoy: A tangible interface to control the emotions of a synthetic character. In: Proc. of AAMAS 23. S. 88 89. 23. [VP99] Vyzas, E. und Picard, R.: Online and offline recognition of emotion expression from physiological data. In: Workshop on Emotion-Based Agent Architectures at the International Conference on Autonomous Agents. Seattle, WA, USA. 999. 64