Bio-Feedback Based Simulator for Mission Critical Training

Similar documents
Active User Affect Recognition and Assistance

Augmented Cognition to enhance human sensory awareness, cognitive functioning and psychic functioning: a research proposal in two phases

Emotion Affective Color Transfer Using Feature Based Facial Expression Recognition

PHYSIOLOGICAL RESEARCH

Affective Game Engines: Motivation & Requirements

Affective Preference From Physiology in Videogames: a Lesson Learned from the TORCS Experiment

Blue Eyes Technology

Beyond AI: Bringing Emotional Intelligence to the Digital

Introduction to affect computing and its applications

VIDEO SURVEILLANCE AND BIOMEDICAL IMAGING Research Activities and Technology Transfer at PAVIS

MODULE 41: THEORIES AND PHYSIOLOGY OF EMOTION

Facial expression recognition with spatiotemporal local descriptors

Choosing Mindray SpO2

Emotion based E-learning System using Physiological Signals. Dr. Jerritta S, Dr. Arun S School of Engineering, Vels University, Chennai

Gender Based Emotion Recognition using Speech Signals: A Review

Determining Emotions via Biometric Software

Improving rapid counter terrorism decision making

N RISCE 2K18 ISSN International Journal of Advance Research and Innovation

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

An assistive application identifying emotional state and executing a methodical healing process for depressive individuals.

COMPUTER PLAY IN EDUCATIONAL THERAPY FOR CHILDREN WITH STUTTERING PROBLEM: HARDWARE SETUP AND INTERVENTION

SMARTfit Multisensory Learning Develops the Brain and the Body. A 30 minute PE class on SMARTfit helps the brain to learn better

Squid: Exercise Effectiveness and. Muscular Activation Tracking

FILTWAM. A Framework for Online Affective Computing in Serious Games. Kiavash Bahreini Wim Westera Rob Nadolski

Smart. Training. Developing advanced exercise machines

Contrastive Analysis on Emotional Cognition of Skeuomorphic and Flat Icon

Fudan University, China

World Journal of Engineering Research and Technology WJERT

FACIAL EXPRESSION RECOGNITION FROM IMAGE SEQUENCES USING SELF-ORGANIZING MAPS

A Possibility for Expressing Multi-Emotion on Robot Faces

Bio-sensing for Emotional Characterization without Word Labels

MSAS: An M-mental health care System for Automatic Stress detection

Personalized HealthCare and Agent Technologies

Hybrid EEG-HEG based Neurofeedback Device

Affective Computing for Intelligent Agents. Introduction to Artificial Intelligence CIS 4930, Spring 2005 Guest Speaker: Cindy Bethel

Promoting Self-Regulation of Students with ASD Using Biofeedback Assistive Technology. Nancy J. Aguinaga, Ph.D. Kristin Bey

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Multimodal Interaction for Users with Autism in a 3D Educational Environment

Smart Sensor Based Human Emotion Recognition

Emotions and Motivation

Potential applications of affective computing in the surveillance work of CCTV operators

INTERACTIVE GAMES USING KINECT 3D SENSOR TECHNOLOGY FOR AUTISTIC CHILDREN THERAPY By Azrulhizam Shapi i Universiti Kebangsaan Malaysia

Artificial Intelligence Lecture 7

Psychology as a Science of Design in Engineering

Neurostyle. Medical Innovation for Better Life

Dexcom Current and Future Technology Innovations. Jorge Valdes Chief Technical Officer

R Jagdeesh Kanan* et al. International Journal of Pharmacy & Technology

A Survey on Serious Games for Rehabilitation

Electronic Care Centre for Caregivers of Patients with Alzheimer's disease

Valence and Gender Effects on Emotion Recognition Following TBI. Cassie Brown Arizona State University

The Leap Motion controller: A view on sign language

Predicting Sleeping Behaviors in Long-Term Studies with Wrist-Worn Sensor Data

ELEC ENG 4BD4 Lecture 1. Biomedical Instrumentation Instructor: Dr. Hubert de Bruin

Team Members: MaryPat Beaufait, Benjamin Dennis, Taylor Milligan, Anirudh Nath, Wes Smith

Mental State Recognition by using Brain Waves

Agents and Environments

Modeling and Implementing an Adaptive Human-Computer Interface Using Passive Biosensors

FitFormula. Sam Liu (Apper) Simran Fitzgerald (Programmer) Jonathan Tomkun (Programmer) ECE1778 Final report

Relationship of Terror Feelings and Physiological Response During Watching Horror Movie

Dimensional Emotion Prediction from Spontaneous Head Gestures for Interaction with Sensitive Artificial Listeners

Affect/Emotion in Design

Exercises 101. Pat Anders Manager, Health Emergency Preparedness Exercises

An Overview of BMIs. Luca Rossini. Workshop on Brain Machine Interfaces for Space Applications

Hand of Hope. For hand rehabilitation. Member of Vincent Medical Holdings Limited

Concepts Weight Market Size Profit Potential Technology Readiness Market Entry Difficulty

Re: ENSC 370 Project Gerbil Functional Specifications

D4.10 Demonstrator 4 EoT Application

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

A USER INDEPENDENT, BIOSIGNAL BASED, EMOTION RECOGNITION METHOD

A Sleeping Monitor for Snoring Detection

INCREASING SIA ARCHITECTURE REALISM BY MODELING AND ADAPTING TO AFFECT AND PERSONALITY

Intelligent Biofeedback using an Immersive Competitive Environment

Modeling Emotion and Temperament on Cognitive Mobile Robots

This is the accepted version of this article. To be published as : This is the author version published as:

PROFIBUS ANALYZER INFOLETTER 14/09

Virtual Dentistry: The Forces Behind the Pain

AC : USABILITY EVALUATION OF A PROBLEM SOLVING ENVIRONMENT FOR AUTOMATED SYSTEM INTEGRATION EDUCA- TION USING EYE-TRACKING

Human Dynamics of Change Part 1. September 9, 2009

A framework for the Recognition of Human Emotion using Soft Computing models

Artificial Intelligence

Finding the Truth: Interview and Interrogation Training Simulations. Ron Punako Senior Software Engineer

Emotion Recognition using a Cauchy Naive Bayes Classifier

Thought Technology Ltd.

Chapter 8. What Is Emotion? What Do Our Emotions Do For Us? Emotion and Motivation

Pride. Jessica L. Tracy. University of British Columbia

by David Haniff Technology Intrusive Thoughts (a Work in Progress) TILT MAGAZINE FALL

Recognising Emotions from Keyboard Stroke Pattern

Emotions and Deception Detection Skills

Navigator: 2 Degree of Freedom Robotics Hand Rehabilitation Device

How-To Evaluate a Veterinary Digital Radiography System A SPECIAL REPORT

Application of Phased Array Radar Theory to Ultrasonic Linear Array Medical Imaging System

Recognizing, Modeling, and Responding to Users Affective States

User Affective State Assessment for HCI Systems

Introduction to Neurofeedback. Penny Papanikolopoulos

A Multimodal Interface for Robot-Children Interaction in Autism Treatment

A BRIEF REVIEW: ASSISTIVE TECHNOLOGY AND AUTISM

Continuing the momentum to deliver improved performance. Robert Cascella Chief Business Leader Diagnosis & Treatment

Real-time Heart Monitoring and ECG Signal Processing

Empathy for Max. (Preliminary Project Report)

The methods we were using to approach our past research

Transcription:

Bio-Feedback Based Simulator for Mission Critical Training Igor Balk Polhemus, 40 Hercules drive, Colchester, VT 05446 +1 802 655 31 59 x301 balk@alum.mit.edu Abstract. The paper address needs for training and simulation tools needed to address a multitude of threats to national security. In order to most effectively prevent terrorist threats, government agencies need to be able to train for extreme event scenarios that are very difficult to realistically replicate. A crucial aspect of the training challenge is to provide scalable, real-time handson training capability that includes not only visual demonstration and description of possible threats but also the capability for the individuals to interactively practice specific threat scenarios and be evaluated on their performance. This paper describes a bio-feedback system incorporated into the trainingsimulator for the mission critical practices. 1 Identification and Significance One of the major goals of human computer interaction (HCI) studies is to develop adaptive, smart computer systems, which are reconfigured based on user response. Such systems use biometric identification such as facial recognition, fingerprint recognition and eye tracking in order to customize access to the system. However, such systems do not adapt to the user need based on their psycho-physiological conditions. This paper describes development of non-invasive touch based system, which will allow tuning a computer based training system to the user emotional and stress levels. The system is based upon the fact that during the operation of a computing device, the user stays in almost permanent manual contact with an input device, such as a mouse, tracking ball, joystick or for more advanced version of the simulators integrated with motion or eye tracking devices. This paper will describe the incorporation of sensors to monitor the physiological conditions, such as galvanic skin response (GSR), electric skin resistance (ESR), and temperature. Based on the sensory response, the computer system would be able to determine stress level and emotional condition of the user, visual-somatic response and adjust itself to provide better training and determine if the user is suitable to the task assigned. P.M.A. Sloot et al. (Eds.): ICCS 2003, LNCS 2659, pp. 1074 1080, 2003. Springer-Verlag Berlin Heidelberg 2003

Bio-Feedback Based Simulator for Mission Critical Training 1075 There are many practical applications for such a system. For example, it could allow parental control of minors, allowing them to play computer games at an emotionally safe level. The described system can enable user interface developers to better monitor the human response to a prototype. Also, the system will enable game developers to dynamically adjust the difficulty level based upon the player s condition. The system can also be integrated into a car steering wheel along with an onboard computer to monitor the motorist s condition during the trip and prevent him/her from falling asleep at the wheel. But one of the most important applications of the system is to monitor the work related stress and enable failure prevention in mission critical tasks in nuclear stations, hazardous chemical, oil factories and other high risk areas. Today s governments face the monumental challenge of addressing a multitude of threats to national security. In order to most effectively prevent terrorist threats, government agencies need to be able to train for extreme event scenarios that are very difficult to realistically replicate. A crucial aspect of the training challenge is the ability to provide a large number individuals amongst numerous agencies with a scalable, real-time hands-on training capability that includes not only visual demonstration and description of possible threats but also the capability for the individuals to interactively practice specific threat scenarios and be evaluated on their performance. At present, such systems do not exist in the market. However, active research is being performed by several groups, such as, IBM BlueEyes team [1, 9], MIT Media Lab [2] and other organizations [1]. The major advantage of the described technique compared to those reported in literature is the monitoring of visual-motor reactions and attention distributions, both very important parameters for training simulator and game developers. For example, the Emotional mouse, the experimental system developed at IBM is limited to measuring six emotional states anger, disgust, fear, happiness, sadness and surprise. This paper describes a system which is significantly more advanced than those being investigated. The system will have a significant impact on the human computer interaction development in the next decade and will lead to more expressive communication with machines. 2 Overview Paul Ekman s facial expression work demonstrated that there is a correlation between a person s psychological state and a person s physiological measurements. Based on this work he developed Facial Action Coding System [3]. Later, Dryer [4] determined how physiological measures could be used to distinguish the various emotional states. He also demonstrated that people consider a computer system as having personality [5], and a system having a similar personality as the user leads to better collaboration [6]. In his experiments Johnson [7] used video cameras to determine users psychological state and concluded that there is a correlation between patterns in user behavior and users psychological state.

1076 I. Balk Wendy Ark [9] in her work reported that during normal computer use (creating and editing documents and surfing the web) people spend about 30% of their total computer time touching their input device. In addition, it has been demonstrated that the major physiological parameters, associated with physiological state, are Galvanic Skin Response (GSR), Electric Skin Resistance (ESR), General Somatic Activity (GSA), temperature (TEM), heart rate (HR) and blood pressure (BP) [3, 4]. Based on the prior art we decided to integrate a bio-sensing device into motion or eye tracking hardware and make it in a way, that it actively monitors the above listed variables GSR, ESR, GSA and HR. 3 Technical Characteristics The major objective was to develop a scanning device, which would be able to monitor the real time psycho-physiological state of the trainee. We term such a system a bio-sensor. The bio-sensor will enable the study of - Psychosomatic and psychotherapeutic issues related to the HCI - Psychological state and personality traits of the subject - Characteristic peculiarities - Mental performance level - Central nervous system state - Functional abilities of central regulation and peripheral blood circulation of cardio-vascular system - Psychological tension and stress level In order to complete these tasks three major objectives need to be achieved (Figure 1). First algorithms to process analog signals from the sensors should be developed. Second, a set of psycho-physiological tests has to be developed to collect and analyze data. And third, an interface should be developed for both report generation and other software integration. Fig. 1. Major modules of the system Each of these three modules is described in more detail below.

Bio-Feedback Based Simulator for Mission Critical Training 1077 3.1 Data Collection Module The data collection module consists of two parts hardware sensors and control software. A data collection module has following characteristics: - Include sensors of photo-plethysmogram and GSR - Tunable discretization frequency up to 1k Hz - Bandwidth of bio-amplifiers not less then 0.1-1000 Hz - Electric shock protection - Power supply from computer Control software properties: - Real time signal filtering - Spectral and autocorrelation analysis algorithms - Bio-amplifiers bandwidth control - Signal intensity control - Automatic detection of photo-plethysmogram parameters 3.2 Data Interpretation Module This is the main analytical module of the system. The goal of this module is to manage data collection and to extract meaningful information from the data. It should provide a set of tests and analysis algorithms, which will study different aspect of HCI. The main functions of the data interpretation module will include: - Chrono-cardiography analysis based of variation principle - Simple and complex visual-semantic response study - Attention distribution test - Associative test - Response speed test - Stress level determination 3.3 Interface Module In order to be a useful tool the system has to be able to generate reports and communicate to other software. This brings us to the third objective to generate an easy to use report generation system and application programmer interface (API). Report generation system should be able to display graphical physiological data and test results as well as being able to save results and provide access to stored user data. The API should be easy to use and allow third party software developers easy access to collected data and diagnostic results.

1078 I. Balk 4 Conclusion and Future Work Development of a smart adaptive computer system has been one of the major challenges of computer science in the last decade. Several different tools and methodologies have been developed in order to obtain some physiological feedback from user. One of the most well known applications of such systems is the polygraph or lie detector which is now widely used in criminology and human resources and cardio-systems in gymnasiums. Several other techniques have been used for biofeedback in psychotherapy. The main concept of this work was to develop an easy to use system, which includes a bio-sensing device and software package integrated with modern training simulator hardware, which will determine users psychological conditions using biofeedback. Data from the IBM research team working on emotional mouse (Table 1-4) shows strong correlation between emotional state and physiological data. The next step that we are planning is to extract not only the emotional data but also the stress level, activity level and attention level. We are planning to integrate different sensors in to the motion or eye tracker hardware. First, we will use infrared detectors in order to obtain photoplethysmogram of the computer user. This signal is depending on blood flow in fingers and allows the detection RR intervals (time between two consecutive heart pulses). Performing statistical analysis and filtering of the RR intervals, we will obtain heart rate and heart rate variability (VF) [10]. A high value of VF indicates that the user is falling asleep and a low value of VF indicates that the user is overloaded or tired. Average RR value for human is 900-1000 milliseconds (equivalent of heart rate 60-70), and average value of VF is 4-5%. Based on this data, the system will be able to warn when the user is too tired or about to fall asleep or getting too stressed to adequately complete the mission. Another analytical technique that we are planning to incorporate into our system is the analysis of video-somatic reaction. This technique is excellent for action based computer game developers. The system will collect reaction time of the user and apply mathematical models similar to the one used to determine a heart rate variability. This will allow us to correlate to the psychological condition of the user. In this case, we will measure average response time and its variability. The advantage of this method is that no hardware modification is required. Combination of heart rate variability, video-somatic reaction based methods and emotional state determination technique similar to one described by IBM [9] will allow error minimization and generate a system which will be able to help make a computer which will be more adjustable to users psychological state

Bio-Feedback Based Simulator for Mission Critical Training 1079 Table 1. Difference Scores. (Source: IBM) GSA GSR Pulse Temp Anger Disgust Fear Happiness Sadness Surprise Mea -0.66-1.15-2.02.22 0.14 -.1.28 Std. 1.87 1.02 0.23 1.60 2.44 1.16 Mea - -53206-61160 -38999-417990 -41242 Std. 63934 8949 47297 46650 586309 24824 Mea 2.56 2.07 3.28 2.40 4.83 2.84 Std. 1.41 2.73 2.10 2.33 2.91 3.18 Mea 1.36 1.79 3.76 1.79 2.89 3.26 Std. 3.75 2.66 3.81 3.72 4.99 0.90 Table 2. Standardized Discriminant Function Coefficients. (Source: IBM) Function 1 2 3 4 GSA 0.593-0.926 0.674 0.033 GSR -0.664 0.957 0.350 0.583 Pulse 1.006 0.484 0.026 0.846 Temp. 1.277 0.405 0.423-0.293 Table 3. Functions at Group Centroids. (Source: IBM) EMOTION Function 1 2 3 4 anger -1.166-0.052-0.108 0.137 fear 1.360 1.704-0.046-0.093 sadness 2.168-0.546-0.096-0.006 disgust -0.048 0.340 0.079 0.184 happiness -0.428-0.184 0.269-0.075 surprise -1.674-0.111-0.247-0.189 Table 4. Classification Results. (Source: IBM) Original Predicted Group Membership Total EMOTION Anger Fear Sadness Disgust Happiness Surprise anger 2 0 0 0 2 1 5 fear 0 2 0 0 0 0 2 sadness 0 0 4 0 1 0 5 disgust 0 1 0 1 1 0 3 happiness 1 0 0 0 5 0 6 surprise 0 0 0 0 1 2 3

1080 I. Balk Acknowledgments.We would like to thank Vladimir Bryskin and the staff of the BioLab Inc. for the help in the work and the information and ideas provided to the authors References 1. http://www.ibm.com 2. http://media.mit.edu 3. Ekman, P. and Rosenberg, E. (Eds.) (1997). What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS). Oxford University Press: New York 4. Dryer, D.C. (1993). Multidimensional and Discriminant Function Analyses of Affective State Data. Stanford University, unpublished manuscript 5. Dryer, D.C. (1999). Getting personal with computers: How to design personalities for agents. Applied Artificial Intelligence, 13, 273-295. 6. Dryer, D.C., and Horowitz, L.M. (1997). When do opposites attract? Interpersonal complementarity versus similarity. Journal of Personality and Social Psychology, 72, 592-603 7. Johnson, R.C. (1999). Computer Program Recognizes Facial Expressions. EE Times www.eetimes.com, April 5, 1999. 8. Picard, R. (1997). Affective Computing. MIT Press: Cambridge 9. Ark, W., Dryer, D.C. and Lu, J. L. Emotional Mouse. IBM White Paper, www.ibm.com 10. European Society of Cardiology and North American Society of Pacing and Elerctrophysiology. Heart Rate Variability. Standards of Measurement, Physiological Interpretation and Clinical Use.