Natural Emotion Expression of a Robot Based on Reinforcer Intensity and Contingency

Similar documents
Introduction to affect computing and its applications

Affective Dialogue Communication System with Emotional Memories for Humanoid Robots

Emotion Explained. Edmund T. Rolls

Emotions and Motivation

Emotions. These aspects are generally stronger in emotional responses than with moods. The duration of emotions tend to be shorter than moods.

Affective Computing for Intelligent Agents. Introduction to Artificial Intelligence CIS 4930, Spring 2005 Guest Speaker: Cindy Bethel

Towards a model of fear in Soar

Emotions of Living Creatures

Practice Question MOTIVATION AND EMOTION. Motivation as Drives. Motivation 10/22/2012

Affective Computing Ana Paiva & João Dias. Lecture 1. Course Presentation

Emotion Lecture 26 1

Artificial Emotions to Assist Social Coordination in HRI

What Are Emotional States, and Why Do We Have Them?

EMOTIONS S E N I O R S P E C I A L I S T I N P S Y C H I A T R Y A N D S E X T H E R A P Y

An Adaptable Fuzzy Emotion Model for Emotion Recognition

PSYC 222 Motivation and Emotions

A ective Computing of Constitutional States for Human Information Interaction

Emotion, higher-order syntactic thoughts, and consciousness

Ch. 18. Brain Mechanisms of Emotion. Bear et al., Neuroscience: Exploring the Brain, 3 rd ed., Lippincott Williams & Wilkins, 2006, pp

Hierarchically Organized Mirroring Processes in Social Cognition: The Functional Neuroanatomy of Empathy

WHAT ARE the COMPONENTS OF THE NERVOUS SYSTEM?

Sociable Robots Peeping into the Human World

Emotion and Motivation. Chapter 8

Affect in Virtual Agents (and Robots) Professor Beste Filiz Yuksel University of San Francisco CS 686/486

The Huggable Robot Probo, a Multi-disciplinary Research Platform

Taken From The Brain Top to Bottom //

9.01 Introduction to Neuroscience Fall 2007

Behavioral and Motivational mechanisms of Brain. Limbic system and the Hypothalamus

Basic Nervous System anatomy. Neurobiology of Happiness

Memory, Attention, and Decision-Making

Basic Brain Structure

Neuroscience Optional Lecture. The limbic system the emotional brain. Emotion, behaviour, motivation, long-term memory, olfaction

Physiology Unit 2 CONSCIOUSNESS, THE BRAIN AND BEHAVIOR

synapse neurotransmitters Extension of a neuron, ending in branching terminal fibers, through which messages pass to other neurons, muscles, or glands

Emotion Theory. Dr. Vijay Kumar

Chapter Eight: Emotion and Motivation

Useful Roles of Emotions in Animated Pedagogical Agents. Ilusca L. L. Menezes IFT6261 :: Winter 2006

Brain Structures. Some scientists divide the brain up into three parts. Hindbrain Midbrain Forebrain

MODULE 41: THEORIES AND PHYSIOLOGY OF EMOTION

Presents. Manage Your Emotions. Arlene R. Taylor PhD Brain References

Motivation and Emotion

Recap: Introduction & History of Motivation & Emotion (Lecture 01 - Ch 1 & 2, Reeve, 2009)

Outline. Emotion. Emotions According to Darwin. Emotions: Information Processing 10/8/2012

Physiology Unit 2 CONSCIOUSNESS, THE BRAIN AND BEHAVIOR

Time-varying affective response for humanoid robots 1

SLHS1402 The Talking Brain

Psych3BN3 Topic 4 Emotion. Bilateral amygdala pathology: Case of S.M. (fig 9.1) S.M. s ratings of emotional intensity of faces (fig 9.

Human Nervous System

1 Introduction. Keywords: Robot emotions, artificial emotion systems, emotions for artificial intelligence.

Predictive Regulation: Allostasis, Behavioural Flexibility and Fear Learning

MEMORY SYSTEMS IN THE BRAIN

Nature of emotion: Six perennial questions

PART I. Neural Foundations of Behavior: The Basic Equipment

Nature of emotion: Six perennial questions

A Formal Model of Emotional-Response, Inspired from Human Cognition and Emotion Systems

A Possibility for Expressing Multi-Emotion on Robot Faces

THE PREFRONTAL CORTEX. Connections. Dorsolateral FrontalCortex (DFPC) Inputs

Human Information Processing and Cultural Diversity. Jaana Holvikivi, DSc. School of ICT

Psychology in Your Life

Motives as Intrinsic Activation for Human-Robot Interaction

Objectives. ! Describe the major structures of the nervous system. ! Explain how a nerve impulse is transmitted.

Lesson 14. The Nervous System. Introduction to Life Processes - SCI 102 1

Lecture 7 Part 2 Crossroads of economics and cognitive science. Xavier Gabaix. March 18, 2004

Design of an Emotion Management System for a Home Robot Andriy Bondarev <ISBN > /ac

The Nervous System. Biological School. Neuroanatomy. How does a Neuron fire? Acetylcholine (ACH) TYPES OF NEUROTRANSMITTERS

Neural Communication. Central Nervous System Peripheral Nervous System. Communication in the Nervous System. 4 Common Components of a Neuron

Prof. Greg Francis 7/31/15

CEREBRUM & CEREBRAL CORTEX

Okami Study Guide: Chapter 2 1

NEURO-BRAIN BOOTCAMP Expanding Leadership and Creativity through the Miracle of Modern (Neuro)Science

Brain Mechanisms of Emotion 1 of 6

INTRODUCTION TO EDUCATIONAL NEUROSCIENCE

Motivation represents the reasons for people's actions, desires, and needs. Typically, this unit is described as a goal

How to build robots that make friends and influence people

The Nervous System and the Endocrine System

Affective Game Engines: Motivation & Requirements

Using progressive adaptability against the complexity of modeling emotionally influenced virtual agents

Where do emotions come from? What are they made of? Why do we have them? ADAPTIVE - Survival. Emotion. AP Psych Myers Ch. 13

Essentials of Human Anatomy & Physiology. Seventh Edition. The Nervous System. Copyright 2003 Pearson Education, Inc. publishing as Benjamin Cummings

Psychology Formative Assessment #2 Answer Key

On the functional contributions of emotion mechanisms to (artificial) cognition and intelligence

Nervous System. Chapter Structure of the Nervous System. Neurons

DEFINING EMOTION 11/19/2009 THE BIOLOGY OF EMOTION & STRESS. A change in physiological arousal, ranging from slight to intense.

Affective Systems. Rotterdam, November 11, 2004

Introduction to Physiological Psychology Psych 260

From Affect Programs to Higher Cognitive Emotions: An Emotion-Based Control Approach

MPEG-4 Facial Expression Synthesis based on Appraisal Theory

Modeling the Mechanisms of Emotion Effects on Cognition

THE NERVOUS SYSTEM CONCEPT 2: THE VERTEBRATE BRAIN IS REGIONALLY SPECIALIZED

Chapter 18: The Brain & Cranial Nerves. Origin of the Brain

Homework Week 2. PreLab 2 HW #2 Synapses (Page 1 in the HW Section)

Modeling Emotion and Temperament on Cognitive Mobile Robots

A context-dependent attention system for a social robot

Psychology in Your Life

Nervous System C H A P T E R 2

Motivation, Conflict, Emotion. Abdul-Monaf Al-Jadiry, MD; FRCPsych Professor of Psychiatry

The Frontal Lobes. Anatomy of the Frontal Lobes. Anatomy of the Frontal Lobes 3/2/2011. Portrait: Losing Frontal-Lobe Functions. Readings: KW Ch.

Brain-Behavior Network. Central Nervous System. Cerebral Cortex Gyrus and Sulcus. Nervous System

A Computational Model of Emotional Conditioning in the Brain

Geography of the Forehead

Transcription:

2007 IEEE International Conference on Robotics and Automation Roma, Italy, 10-14 April 2007 ThB4.3 Natural Emotion Expression of a Robot Based on Reinforcer Intensity and Contingency Seung-Ik Lee, Gunn-Yong Park, and Joong-Bae Kim Abstract An emotional robot is regarded as being able to express its diverse emotions in response to internal or external events. This paper presents a robot affective system that is able to express life-like emotions. In order to do that, the overall architecture of our affective system is based on neuroscience from which we obtained the natural emotional processing routines. Based on that architecture, we apply the reinforcer effects expecting that those would lead the affective system to be more similar to real-life s emotion expression. The robot affective system has responsibility for gathering environmental information and evaluating which environmental stimuli are rewarding or punishing. The emotion processing involves with appraisal of the external and internal stimuli, such as homeostasis, and generates the affective states of the robot. Therefore, emotions are associated with the presentation, omission, and termination of the expected rewards or punishers (reinforcers). The experimental results show that our affective system can express several emotions simultaneously as well as the emotions decrease, increase, or changes to another emotion seamlessly as time passes. I. INTRODUCTION An emotion is an essential part of the human-robot interaction (HRI), since emotions influence rational decisionmaking, perception, learning, and other cognitive functions of a human [1]. According to the somatic marker hypothesis [2], the marker records emotional reaction to a situation. We learn the markers throughout our lives, and use them for our decision-making. Therefore, an emotional robot needs an affective system to synthesize and express emotions. Due to the importance of the emotion in the HRI, there have been several efforts to create robots with emotions to improve the HRI, such as the Sony Aibo [3], and the MIT Kismet [4]. Much of the previous work to build emotional robots is inspired by ethology, and use an affect space to produce an emotion. However, the emotional robot based on the affect space expresses only one emotion at a time, because the affect space has a competitive relationship between emotions. For example, Aibo has six emotions based on Takanishi s model [5], and always expresses only one affective state from among its six emotions. However, contrary to the previous This work was supported in part by MIC & IITA through IT Leading R&D Support Project. Authors are with Intelligent Robot Research Division, Electronics and Telecommunications Research Institute, 161 Gajeong-dong, Yuseong-gu, Daejeon, 305-700, KOREA (e-mail: {the silee, pgy64257, jjkim}@etri.re.kr) work, humans can have several emotions simultaneously and can express them in various ways. Because of the dynamicity of the human emotional processing, there have been several attempts to understand the human emotion processing in the view of neuroscience and cognitive science. Cognitive science studies the computational and representational structure of the mind by combining psychology, artificial intelligence, philosophy, and linguistics. In the cognitive science, an emotion involves with external events. For example, the Ortony, Clore, and Collins (OCC) model [6] assumes that emotions arise from valenced reactions to the phenomena perceived in the environment. These reactions to situations consist of a success (or failure) of goals, an attractiveness of objects, and a praiseworthiness (or blameworthiness) of actions. However, because of weak points of cognitive model [7], we suggest a neuroscientific approach to overcome these difficulties. Contrary to cognitive science, neuroscience is a biological study of the human brain that deals with the structure, function, development, genetics, biochemistry, and pathology of the nervous system. According to neuroscientific researches, human has a complicate and dynamic emotion processing mechanism due to the considerable development in the human brain. Therefore, we can generate and express various affective states simultaneously. Furthermore, because people feel more comfortable with a human-like robot [1], human-oriented robot should have an emotion generation method similar to that of a human. Hence, we propose a robot affective system for intelligence robots inspired from neuroscientific research. We proposed neuroscientific affective system to overcome disadvantages of previous works such as affective space model and cognitive emotion model. The emotional expression of the affective space model is limited because this model always generates only one emotion at a time. In addition, the cognitive emotion model needs to be extended with a history function and a personality [7] because it overlooks internal stimuli such as homeostasis. However, contrary to the affective space model, the proposed system generates various emotional expressions and implements internal stimuli such as homeostasis. This paper is organized as follows. Section II presents the human emotion processing in the aspect of both neuroscience 1-4244-0602-1/07/$20.00 2007 IEEE. 2144

C D E A: Somatosensory Cortex B: Hypothalamus C: Orbitofrontal Cortex D: Olfactory Cortex External Stimuli F B A G H E: Primary Taste Cortex F: Amygdala G: Inferior Temporal Visual Cortex H: Posterior Inferior Temporal Visual Cortex (a) Lateral view of the macaque monkey brain Appraisal Process Em otion Process Expression Primary Reinforcers Secondary Reinforcers Amygdala Hypothalamus Orbitofrontal Cortex (b) Diagram of the macaque monkey brain Fig. 1. Brain of the macaque monkey Celebral Cortex and cognitive science. Section III describes the proposed robot affective system based on Section II. The experimental result and the conclusion are presented in Section IV and Section V, respectively. II. AN AFFECTIVE SYSTEM IN A HUMAN BRAIN Since the temporal lobe and the prefrontal cortex have undergone considerable development [8], [9], humans can feel several emotions simultaneously, and have various ways of expressing them. For example, when you are competing with your best friends for a contest, and finally you get the first prize. Then you can feel both joy and grief because of your success as well as the failure of your friend. There have been several efforts to discover the emotion processing mechanism in humans to date. In this section, we introduce a neuroscientific approach to human affective system. Since the triune brain model [1] had been proposed, many people assumed that the limbic system is the source of emotions. However, according to the recent neuroscientific studies [8], [9], [10], the emotion involves not only with the limbic system but also with the neocortex such as the orbitofrontal cortex. The amygdala and the orbitofrontal cortex are responsible for perception, appraisal, and learning of environmental stimuli such as visual or acoustic sensory stimuli. The hypothalamus is in charge of homeostasis, i.e., maintaining the body s status quo. It controls body temperature, hunger, thirst, and other circadian cycles, and alters the emotion to obtain the desired stimuli. Fig. 1 (a) shows how the primate brain responds to the environmental stimuli. In primates, the external stimuli are decoded as a reward or a punishment. Rewards are stimuli for which an animal will work, and punishers are stimuli that the animal will work to avoid. For example, in the primate taste system, the reward decoding occurs after several processing stages. The first stage is the primary taste cortex, which identifies the taste. According to that result, the secondary taste cortex in orbitofrontal cortex evaluates the reward value. This value is decreased by feeding to satiety, because neurons in the orbitofrontal cortex decrease their responses as the reward value of the food decreases. Since an animal performs any actions to obtain the reward or to avoid the punisher [8], rewards and punishers can be defined as reinforcers. As shown in Fig. 1 (b), there are two types of reinforcers: primary reinforcers, and secondary reinforcers [9]. The primary reinforcer (e.g., the taste of food or pain) is unlearned, and has innate reinforcing properties. The secondary reinforcer (e.g., the auditory or visual stimulus) develops its reinforcing properties through learning its association with primary reinforcement. For example, fear might be produced by a sound (i.e., the conditioned stimulus) that has been previously associated with a painful stimulus (i.e., the primary reinforcer). This type of learning is referred to stimulus-reinforcement association learning, [9] and involves with the amygdala and the orbitofrontal cortex. For example, the association learning in the amygdala is realized as modifiable synapses between visual or auditory neurons and neurons from taste, olfactory or somatosensory primary reinforcers [9]. Emotions are defined as affective states produced by rewards and punishers [9]. For example, a pleasure could be produced by an occurrence of the rewards such as an admiration of others or a pleasant touch; and frustration or anger could be produced by the termination of a reward such as the death of a loved one. As shown in these examples, the affective states involve the occurrence, termination, or omission of the rewards (or punishers). As a result, Fig. 2 shows that some emotions are associated with both the contingency of reinforcer and its intensity [8]. The affective states are classified as primary and secondary emotions. The primary emotions are innate, and involve the primary reinforcers. For example, when you are surprised by a loud noise, the surprise is related to the primary reinforcer such as an acoustic sensory stimulus. On the other hand, the secondary emotions involve the secondary reinforcers, and depend on results of the stimulus-reinforcement association 2145

Presentation of Reward Ecstasy Robot Affective System Omission or Termination of Reward Rage Anger Frustration Grief Sadness Elation Pleasure Apprehension Fear Relief Omission or Termination of Punisher Environment Sensor Module Actuator Module Logical Sensor Data Behavior Emotion Processing Module Emotions Expression Module Appraisal References Behaviors Long-Term Memory Terror Presentation of Punisher Fig. 3. Framework for the robot affective system Fig. 2. Emotions associated with reinforcement contingencies learning. The human emotion processing involves with both reinforcers and current states of neuromodulation [11]. Neuromodulation refers to the action of a neuromodulator on nerve cells. The neuromodulator, such as dopamine, serotonin, and opioids, is released by a neuron and end up spending a significant amount of time in the cerebrospinal fluid and modulating the overall activity level of the brain. Thus, a neuromodulator can change the way of computation between single neurons and synapses, either enhancing or damping their activities. As described above, the environmental stimuli are evaluated in terms of the primary or secondary reinforcers including the neuromodulators, and then the affective states are dynamically generated by the contingencies and intensity of these reinforcers. Hence, human can have several emotions simultaneously and express them in various ways. Therefore, we propose a dynamic affective system inspired from neuroscientific researches for a robot to be more human-like and friendly. III. THE ROBOT AFFECTIVE SYSTEM For a robot to have more natural emotion, we propose a robot affective system inspired from neuroscience such that it can have various emotional states at the same time and express those combined emotions just like humans do. Our affective system is shown in Fig. 3. The sensor module translates external stimuli into logical sensor information, similar to the orbitofrontal cortex or primary taste cortex. The logical sensor data is an abstraction of the external stimuli (e.g., a name of an object). The actuator module realizes emotional behaviors of a robot. The robot affective system is responsible for generating and expressing affective states according to the external or internal stimuli. The robot affective system comprises Logical Sensor Data Behavior Fig. 4. Appraisal Unit Executed Behavior List Long-Term Memory Praiseworthy Action List Desired Object List Emotion Processing Module Appraisal Result Emotion Generation Unit Primary Emotion Joy Sadness Surprise Secondary Emotion Tired Fear Motivation Anger Disgust Social- Interaction Emotions Emotion processing module of the robot affective system three modules: an emotion-processing module, an emotional expression module, and a long-term memory. As amygdala and hypothalamus generates emotions in accordance with external or internal stimuli, the emotion-processing module determines which stimuli are rewards or punishers and generates affective states. The emotional expression module controls the sequential or parallel execution of emotional behaviors such as the facial expressions or text-to-speech of the robot. Lastly, the long-term memory contains the information for the emotion-processing module and the emotional expression module. A. Emotion Processing Module As shown in Fig. 4, the emotion-processing module comprises an appraisal unit, an emotion generation unit, and an external long-term memory. The long-term memory contains a praiseworthy action list and a desired object list for the appraisal unit. The praiseworthy action list and the desired object list are an appraisal reference for the praiseworthiness of the robot action and the attractiveness of the recognized 2146

Energy Level An(t) Energy Level An(t) Catetory TABLE I APPRAISAL RESULT OF APPRAISAL UNIT Appraisal result Positive Negative Recognized object Like Dislike Robot action Praiseworthy Blameworthy Facial expression of a user Good Bad Consequence of a robot action Success Fail Time (a) Appraisal energy level in a similar mood Time (b) Appraisal energy level in a different mood Fig. 5. Characteristic graph of moods object, respectively. 1) Appraisal Unit: The appraisal unit assesses external stimuli (e.g., logical sensor data and emotional behaviors) in accordance with the long-term memory. As shown in TABLE I, the external and the internal stimuli are classified into four categories in the appraisal unit. When a recognized object is in the desired object list; or an action of the robot itself is in the praiseworthy action list; or the facial expressions of a user is good, the appraisal result is positive. When the robot executes an emotional behavior, it is included in the executed behavior list of the appraisal unit. If the listed behavior receives its expected result within the specified time interval, it is a positive appraisal result. In the appraisal unit, the human neuromodulatory system is implemented as moods. Thus, the energy level of each appraisal result A n (t) depends on time and mood, as shown in Fig. 5. In a good mood, the energy level of the positive appraisal result increases more rapidly than that of the negative one, and vice versa. The mood Mood(t) is defined as (1). Mood(t) = A k=1 e k E k (t 1) (1) where E k (t) is the activation level of the k th emotion, A is the number of the emotion, and e k is a gain parameter of the k th emotion. If the activation level E k (t) is the positive emotion, then e k is defined as positive value, and vice verse. When the value of mood Mood(t) is positive, the mood of the system is positive. 2) Emotion Generation Unit: The emotion generation unit comprises three parts: the primary emotion part, the secondary emotion part, and the motivation part (see Fig. 4). The primary emotions involve the primary reinforcers. They are linked directly to the logical sensor data, since the primary reinforcers are unlearned stimuli. On the other hand, the secondary emotions involve the secondary reinforcers. Therefore, they depend on the results from the appraisal unit. The motivation part implements homeostatic regulations in the hypothalamus of the human brain, such as tired and social interaction. According to a psychological research on emotion [12], we express our elemental emotions in the same ways, and can recognize these emotions in the face of another: anger, disgust, fear, joy, sorrow, and surprise. Thus, we propose that the affective state of the robot is the blending of the six basic emotions. In the emotion generation unit, the activation level of each emotion is continuously updated. When the activation level exceeds a threshold value, the emotion is activated. All emotions above the thresholds are active in contrast to the winner-take-all scheme that allows only one emotion to be active at a time. The activation level of the i th primary emotion E i (t) at time t depends on the level of logical sensor data as in (2). E i (t) = E i (t 1)+ M m=1 w m S m (t) δ(t) (2) where S m is the level of the m th logical sensor data, w m is the weight of the mth sensor data, δ(t) is the decay factor of the emotion, and M is the number of logical sensors related to the emotion. For example, if the emotion surprise depends on the microphone level, the robot is surprised by a loud noise. In this system, the elements of the surprise are a loud noise and an abrupt change in the visual stimulus. On the other hand, the secondary emotions depend on the appraisal result of the appraisal unit, because it is related to the stimulus-reinforcement association learning of the secondary reinforcers. The activation level of the j th secondary emotion at time t, E j (t), is affected by the energy level of appraisal result A n (t) and that of the motivation M l (t) as in (3). E j (t) = E j (t 1)+ N n=1 a n A n (t)+ L l=1 c l M l (t) δ(t) (3) where a n is the weight of the n th appraisal result, c l is the 2147

weight of the l th motivation, N is the number of related appraisal results, and L is the number of related motivations. The motivation unit implements homeostatic factors, such as tired for taking a rest and social interaction for interacting with people. In the emotion-generation process, as shown in (3), the activation level of the emotion E j (t) is affected by the activation level of the motivation M l (t). The motivation level M l (t) increases gradually when internal or external stimuli do not satisfy the desired stimuli of each motivation. As shown in (3), the motivation controls the emotion to execute appropriate actions, and consequently tries to obtain the required stimuli. B. Expression Module The emotional expression module controls the emotional actions of the robot. It determines appropriate emotional behavior using the long-term memory to express the affective states. The long-term memory contains a set of emotional behaviors of which there are two types: concurrently executable behaviors, and sequentially executable behaviors. An execution of a sequential behavior depends on which activation level of the emotion is the highest. The execution of sequential behavior related to the highest activation level has priority over the others. In the concurrently executable behaviors, all behaviors that exceed the thresholds are performed. For example, the eye lid position of the robot D EyeLid (t) is defined as in (4). D EyeLid (t) = A a=1 µ a E a (t) P a (t) (4) where A is the number of emotions, P a (t) is the eyelid position designated for a th emotion, E a (t) is the activation level of the a th emotion, and µ a is the weight of P a (t). The emotional expression module should deal with competing rewards, goals, and priorities. Furthermore, the selection process between the emotional behaviors can respond to many different types of reward. Therefore, the emotional behavior has properties, such as the expected result, if any; the processing type (e.g., sequential execution or parallel execution); and the waiting time. The expected result is the available stimulus from the emotional behavior, and the waiting time is the maximum time allowed to receive the expected stimulus from the environment. IV. EXPERIMENTAL RESULTS To evaluate the proposed system, it was applied to the Philips icat, which is an experimentation platform for the HRI. The icat has four touch sensors, two microphones, a CCD camera, and 13 servos for facial postures. The robot affective system runs on a 3 GHz PC running Windows, and communicates with the icat through a USB port. The robot affective system produces the robot s emotional states (a) Calm (b) Joy (c) Sadness (d) Anger (e) Disgust (f) Fear (g) Surprise (h) Tired Data category Fig. 6. Desired object list Praiseworthy action list behavior The basis facial postures of the icat TABLE II DATA STORED IN THE LONG-TERM MEMORY Data Apple, bananas, red-color, toy Make laugh, smiling, nod 8 basis facial posture, exclamations, wink in accordance with the internal stimuli; and external stimuli such as interactions between the user and the icat (e.g., the touch sensors of the icat). The icat s basis facial postures, based on its emotional states, are shown in Fig 6 and stored in the long-term memory. As shown in TABLE II, the long-term memory contains the praiseworthy action list, the desired object list, and the emotional behaviors such as eight basis facial postures. The robot s facial expression is defined as a blending of basis facial postures (see Fig. 7). For example, when the robot likes red color and dislikes blue color, the robot detects a blue object at the beginning of (2) in Fig. 7 (a), and then the activation level of disgust increases. The facial expression of disgust is also shown in (2) of Fig. 7 (b). As shown in (3) of Fig. 7 (a), the joy is activated because of the recognition of a red object. The facial expression in (3) of Fig. 7 (b) is a blending of two basis facial postures (joy and disgust). As shown in (4) of Fig. 7 (a), when the blue object has disappeared, the activation level of disgust decreases, and the emotional expression is in (4) of Fig. 7 (b). Fig. 8 describes an emotion processing associated with the reinforcement contingencies. As shown in Fig. 2, the termination (or omission) of rewards involves with the grief or anger in the human emotion processing. Similarly, the emotion processing in the proposed system is associated with a presentation as well as termination (or omission) of reinforcement. Therefore, when an attractive object is 2148

(a) Activation level of joy and disgust (a) Activation level of joy and sadness (b) Sequence of facial postures in accordance with (a) Fig. 7. Activation level of emotions and facial postures Fig. 8. (b) Sequence of facial postures in accordance with (a) Emotion associated with the reward contingencies unexpectedly disappeared at the end of (2) in Fig. 8 (a), the terminated rewards affects the level of sadness. As described above, the facial expression of the icat is a blending of basis facial postures. In this system, in accordance with the emotion activation level, the emotional expression is the sequential execution, or a linear combination of the parallel behaviors. In contrast to previous work, the proposed system is able to express various emotional behaviors continuously. According to the affective states of the robot, the robot generates various combinational behaviors, such as facial expressions, and sequential behaviors, such as emotional exclamations and simple text-to-speech. V. CONCLUSION We have presented a robot affective system inspired from neuroscience with reinforcer intensity and contingency focused. The proposed robot affective system produces simultaneously various emotions in accordance with the appraisal of external stimuli as well as the contingencies of reinforcement. In addition, emotions decrease, increase, or changes to another emotion seamlessly as time pass, what we believe more real-life like. The facial expression of the system is not a predefined emotional behavior, but the blending of concurrently or sequentially executable behaviors. In future, we intend to implement personality and association learning into the affective system. Personality has an important role in emotion processing and can change the emotional reaction to the environment. Therefore, emotional robots should have characteristic personality. Furthermore, we will also focus on the association learning for the emotional response of unknown objects. REFERENCES [1] R. W. Picard, Affective Computing. MIT PressA, 1997. [2] A. Damasio, Descarte s error: Emotion, reason and the human brain. NewYork, NY: Harper Collins, 1994. [3] R. C. Arkin, M. Fujita, T. Takagi, and R. Hasegawa, An ethological and emotional basis for human-robot interaction, in Proc of the IEEE/RSJ Int. Conference on Intelligent Robots and Systems (IROS2002), 2002. [4] C. Breazeal, Sociable machines: Expressive social exchange between humans and robots, Ph.D. dissertation, Dept. Elect. Eng. And Computer. Sci., MIT, Cambridge, 2000. [5] A. Ortony, G. L. Clore, and A. Collins, The Cognitive Structure of Emotions. Cambridge University Press, 1988. [6] A. Takanishi, An anthropromorphic robot head having autonomous facial expression function for natural communication with human, in Proc of 9th International Symposium of Robotics Research, 1999, pp. 197 304. [7] C. Bartneck, Integrating the occ model of emotions in embodied characters, in Proc of the Workshop on Virtual Conversational Characters: Applications, Methods, and Research Challenges, 2002. [8] E. T. Rolls, The neural basis of emotion, in International Encyclopedia of the Social and Behavioral Sciences, N. J. Smelsner and P. B. Baltes, Eds. Amsterdam, The Netherlands: Pergamon, 2002, pp. 4444 4449. [9], Vision, emotion and memory: from neurophysiology to computation, in Cognition and Emotion in the Brain, N. J. Smelsner and P. B. Baltes, Eds. Amsterdam, The Netherlands: Elsevier Science, 2003, pp. 547 573. [10] S. Greenfield, Brain Story. BBC World Wide Publishing, 2000. [11] M. A. Arbib and J. M. Fellous, Emotions: From brain to robot, in Trends in Cognitive Science. Elsevier Science, 2004. [12] P. Ekman, Are there basic emotions? in Psychological Review, 1992, pp. 550 553. 2149