Katsunari Shibata and Tomohiko Kawano

Similar documents
A model to explain the emergence of reward expectancy neurons using reinforcement learning and neural network $

Learning Utility for Behavior Acquisition and Intention Inference of Other Agent

Realization of Visual Representation Task on a Humanoid Robot

Learning and Adaptive Behavior, Part II

COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks

Exploration and Exploitation in Reinforcement Learning

Sparse Coding in Sparse Winner Networks

Reinforcement Learning of Hierarchical Skills on the Sony Aibo robot

Pupil Dilation as an Indicator of Cognitive Workload in Human-Computer Interaction

Artificial Intelligence Lecture 7

Lecture 13: Finding optimal treatment policies

Integration of a Fuzzy ART Approach in a Biologically Inspired Sensorimotor Architecture

Intelligent Machines That Act Rationally. Hang Li Toutiao AI Lab

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018

Hebbian Plasticity for Improving Perceptual Decisions

Validating the Visual Saliency Model

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

DB HAMMER CURL: 1-LEG SUPPORTED ALT- ARM + ISO-HOLD

Grounding Ontologies in the External World

Observational Learning Based on Models of Overlapping Pathways

Evolution of Plastic Sensory-motor Coupling and Dynamic Categorization

Artificial Intelligence. Intelligent Agents

Reactive agents and perceptual ambiguity

Robot Trajectory Prediction and Recognition based on a Computational Mirror Neurons Model

A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning

ANN predicts locoregional control using molecular marker profiles of. Head and Neck squamous cell carcinoma

Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Agents and Environments

Biologically-Inspired Human Motion Detection

Neurally Inspired Mechanisms for the Dynamic Visual Attention Map Generation Task

Intelligent Machines That Act Rationally. Hang Li Bytedance AI Lab

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

about Eat Stop Eat is that there is the equivalent of two days a week where you don t have to worry about what you eat.

A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences

Strength Training for Marathoners

CS 771 Artificial Intelligence. Intelligent Agents

An Autonomous Distal Reward Learning Architecture for Embodied Agents

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Clinical Examples as Non-uniform Learning and Testing Sets

CSE Introduction to High-Perfomance Deep Learning ImageNet & VGG. Jihyung Kil

Object recognition and hierarchical computation

Design and Implementation study of Remote Home Rehabilitation Training Operating System based on Internet

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

Does scene context always facilitate retrieval of visual object representations?

Things you need to know about the Normal Distribution. How to use your statistical calculator to calculate The mean The SD of a set of data points.

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

Competition Between Objective and Novelty Search on a Deceptive Task

Computational Intelligence Lecture 21: Integrating Fuzzy Systems and Neural Networks

Effect of Sensor Fusion for Recognition of Emotional States Using Voice, Face Image and Thermal Image of Face

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention

arxiv: v1 [cs.ai] 28 Nov 2017

Sensation vs. Perception

Foundations of Artificial Intelligence

22c:145 Artificial Intelligence

CS343: Artificial Intelligence

Neuromorphic convolutional recurrent neural network for road safety or safety near the road

Feature Integration Theory

Discrimination and Generalization in Pattern Categorization: A Case for Elemental Associative Learning

Dynamics of Color Category Formation and Boundaries

Learning the Meaning of Neural Spikes Through Sensory-Invariance Driven Action

Dynamic Rule-based Agent

Learning to Use Episodic Memory

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Cross-modal body representation based on visual attention by saliency

(Visual) Attention. October 3, PSY Visual Attention 1

Competing Frameworks in Perception

Competing Frameworks in Perception

Modeling the Deployment of Spatial Attention

Visual Categorization: How the Monkey Brain Does It

Vorlesung Grundlagen der Künstlichen Intelligenz

Early Learning vs Early Variability 1.5 r = p = Early Learning r = p = e 005. Early Learning 0.

Reinforcement Learning : Theory and Practice - Programming Assignment 1

Re: ENSC 370 Project Gerbil Functional Specifications

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

Category-Based Intrinsic Motivation

A Modular Hierarchical Behavior-Based Architecture

Intelligent Agents. Russell and Norvig: Chapter 2

Changing expectations about speed alters perceived motion direction

AUTONOMOUS LEARNING OF THE SEMANTICS OF INTERNAL SENSORY STATES BASED ON MOTOR EXPLORATION


Computational Cognitive Science

Lesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS

International Journal of Scientific & Engineering Research Volume 4, Issue 2, February ISSN THINKING CIRCUIT

Computational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention.

Subliminal Messages: How Do They Work?

CHAPTER ONE CORRELATION

Stretching. Knees: Rotate your knees in a circle, keeping them together and bending down slightly.

Ultrasound Knobology

Dopamine neurons activity in a multi-choice task: reward prediction error or value function?

12.1 Inference for Linear Regression. Introduction

Learning in neural networks

Artificial Intelligence

19th AWCBR (Australian Winter Conference on Brain Research), 2001, Queenstown, AU

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

Knowledge Based Systems

Computational Neuroscience. Instructor: Odelia Schwartz

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA


Transcription:

Learning of Action Generation from Raw Camera Images in a Real-World-Like Environment by Simple Coupling of Reinforcement Learning and a Neural Network Katsunari Shibata and Tomohiko Kawano Oita University, 700 Dannoharu, Oita, Japan shibata@cc.oita-u.ac.jp http://shws.cc.oita-u.ac.jp/~shibata/home.html Abstract. For the development of human-like intelligent robots, we have asserted the significance to introduce a general and autonomous learning system in which one neural network simply connects from sensors to actuators, and which is trained by reinforcement learning. However, it has not been believed yet that such a simple learning system actually works in the real world. In this paper, we show that without giving any prior knowledge about image processing or task, a robot could learn to approach and kiss another robot appropriately from the inputs of 6240 color visual signals in a real-world-like environment where light conditions, backgrounds, and the orientations of and distances to the target robot varied. Hidden representations that seem useful to detect the target were found. We position this work as the first step towards taking applications of the simple learning system away from toy problems. 1 Introduction In order to develop human-like intelligent robots, researchers have been trying to introduce sophisticated human functions into them. They have modularized the whole process into some functional modules such as recognition, action planning, and control at first, then developed each module individually, and finally connected them sequentially. However, the brain is massively parallel and cohesively flexible, and much of the brain functions seem to be performed subconsciously. Therefore, we think there is a limitation for humans to know exactly how the brain really works by the guess of the functions through the sequential consciousness. For the optimization and consistency of the entire system, we think that even though the understandability for humans is sacrificed, we should try not to interfere to the robot, but to leave everything to the robot s optimization through autonomous learning. From these discussions, we have suggested a general and autonomous learning system in which a neural network (NN) simply connects from sensors to actuators in a robot and is trained by reinforcement learning (RL) as shown in Fig. 1 [1][2]. M. Köppen et al. (Eds.): ICONIP 2008, Part I, LNCS 5506, pp. 754 761, 2009. c Springer-Verlag Berlin Heidelberg 2009

Learning of Action Generation from Raw Camera Images 755 Environment sensor Optimization on a parallel and cohesively flexible system (a) actuator Emergence! sensor recognition carrot stick Neural Reinforcement Network Learning memory planning actuator control attention logical thinking (b) exploration Fig. 1. (a) The objective in robots. (b) Parallel and cohesively flexible learning of the entire process from sensors to actuators by coupling of RL and NN. In our previous work[3], acquisition of flexible image recognition by coupling of RL and a NN was shown, but the robot was static at the same place and could move its head to one of only nine discrete states. This means that the target recognition could be achieved easily by a template matching technique, and actually, in the hidden layers, many neurons seemed to work as a template. In this experiment, since one robot was placed at a different initial location at each trial (episode) and walked, the number of possible views of the other robot could not be counted. The robot has to find the other robot in the camera images under various conditions as shown in Fig. 4, and has to approach and kiss it. One can understand that it is not so easy to develop appropriate programs for this task in the real-world-like environment. To learn a task with such a huge state space, the point is whether effective abstraction could be obtained autonomously through learning. We position this work as the first step towards taking applications of the simple learning system away from toy problems. 2 Experiment Here an experiment using two AIBO robots, a white one and a black one, is introduced. The two AIBOs are placed in a 1.5m 1.6m field. The mission of the black AIBO is to approach the white AIBO and kiss it as shown in Fig. 2. Before each episode, the black AIBO is located manually within the range of about ±45 degrees from the front line of the white AIBO and on the condition that the white AIBO is caught in the camera of the black AIBO. The black AIBO captures a color image with its camera at its nose after moving every step. The number of pixels is around 350,000, but when the AIBO sends the image to the computer, it is reduced to 52 40 = 2080 by pixel skipping. The white AIBO sits on a transparent plastic bottle, and is fixed at the same place during one episode. However, the location of the white AIBO was sometimes changed between episodes.

756 K. Shibata and T. Kawano Start Goal! Fig. 2. The mission of the black AIBO is to walk and kiss the white AIBO Red Pixels Green Pixels Q-value Turn Right Camera Image Blue Pixels Go Straight Turn Left 52x40x3 = 6240 Go Straight Turn Right Turn Left Fig. 3. The processing in the black AIBO. The visual signals are inputted to the NN directly and the black AIBO chooses its action according to the outputs from three possible actions, turn right, turn left and go forward. The processing in the black AIBO is shown in Fig. 3. The 52 40 image is inputted into the 5-layer NN after inverting and normalizing each pixel value between 0.0 and 1.0. Since each pixel is represented by the gradation of RGB colors, the number of inputs to the NN is 6240 in total. The number of the output neurons is three in accordance with the number of actions that the black AIBO can choose. The three actions are turn right, turn left and go forward. The output is dealt with as the Q-value for the corresponding action, and ɛ- greedy (ɛ =0.13 here) is employed as the action selection[4]. The actions are based on the software provided by Ito and Kakiuchi[5]. When the robot chooses the action turn, the turning radius and angle are 22.5cm and 9.5 respectively. When choosing the action go forward, it advances by 6cm. However, the angles and distances actually vary at each time, since it is a real robot.

Learning of Action Generation from Raw Camera Images 757 Different light conditions Different backgrounds Different orientations Different distances Fig. 4. Some sample images captured under different light conditions with different backgrounds, different orientations of and different distances to the white AIBO. They were used actually in learning. An experimenter always monitored the situation and the camera image during the experiment. When it judged that the black AIBO kissed the white AIBO, a reward was given to the black AIBO. When it judged that the black AIBO lost sight of the white AIBO, a penalty was given. In both cases, the episode terminated. For acceleration of learning, some small rewards or penalties were imposed depending on the situation without termination of the episode, but they were not imposed after 2000 episodes of learning. In order to show the variety of the captured images depending on the light condition, background, and orientation of and distance to the white AIBO, Fig. 4 shows some samples. The light conditions change according to the hour when the experiment was done. The location of the white AIBO and also the backgrounds were changed by the experimenter. The black AIBO not only approach from the front of the white AIBO, but also was located diagonally to the front of the white AIBO as an initial state though it was not located at its side or back. As for the NN, the number of neurons in each layer is 6240-600-150-40-3 from input layer to output layer. The training signal is generated using Q-learning[6] and the NN is trained by Error Back Propagation[7]. When one of the actions is chosen at time t, the robot walked, and forward computation of NN at time t+1 is performed with the new camera image S t+1 captured after the walk. After that, the previous input S t are inputted into the NN again, and the training signal is provided only to the output corresponding to the executed action a t at time t according to T at,t = r t+1 + γ max O a (S t+1 ) (1) a where γ is a discount factor (here 0.9 is used), O a (S t )isthea-th output of the NN when the captured image at time t are entered as sensor signal inputs. Here, the sigmoid function used as the output function of each neuron ranges from -0.5 to 0.5. To adjust the offset between Q-value and NN output, a linear transformation is done. Here, since Q-value 0.0 corresponds to output -0.4 and Q-value 0.8 corresponds to output 0.4, 0.4 is added or subtracted actually for the transformation. When the episode terminates, the second term of right-hand

758 K. Shibata and T. Kawano (I) (II) (III) (IV) (a) (V) success ratio[%] 100 80 Background 2 Background 1 60 40 20 0 0 1000 2000 3000 4000 (c) trials Background 1 Background 2 (I) (II) (III) (IV) (V) (b) Fig. 5. (a) The environment for the test trials and (b) the camera images at the five initial locations from (I) to (V) for two kinds of backgrounds. (c) The learning curve shows the change of success ratio according to the number of trials. side of Eq. (1) is set to be 0. The reward for the kiss is 0.8, while the training signal for the missing is equal to the Q-value output minus 0.02. The reward and penalty at non-terminal state is 0.04 and -0.04 respectively. Since each episode takes a long time, simulation was also run unsupervised. All the visual signals and chosen actions were stored for every episode during the experiment, and in the simulation, the NN was trained by using the stored data up to that time. However, since actions could not be chosen, the same episodes appear. The simulation was run 13 times during the course of 4000 experimental episodes. In each simulation, 10000 episodes of learning were executed. Fig. 5(c) shows the learning curve. All the connection weights in the NN are stored every 250 episodes of learning. Two backgrounds were prepared for the test, and five initial locations were selected as shown in Fig. 5(a). A sample image from each initial state is also shown in Fig. 5(b). Five trials were performed for each background and initial location set with the greedy action selection, and the success ratio for each background was observed as a learning curve. Using the connection weight set before learning, it could not kiss the white AIBO from any initial positions in either background. The ratio increased gradually at first, and did not increase very much after around 1000 trials. However, eventually, the success ratio reached between 78% to 90% even though the black AIBO was not exposed to the same background in the learning phase. In order to examine how learning progressed, the time-series of Q-values are observed. Fig. 6 shows the change of Q-values for the two cases of using the connection weight set (a) after 1000 trials and (b)(c) after 4000 trials. If learning converged successfully, it is expected that the maximum Q-values at each step

Learning of Action Generation from Raw Camera Images 759 Q_value Forward Left Right MaxQ IdealQ F: Move forward R: Turn right L: Turn left step 1 2 3 4 5 6 7 8 9 101112 1 2 3 4 5 6 7 1 2 3 4 5 6 7 8 9 101112 0.8 0.8 0.8 0.5 0.5 0.5 Ideal Q = 0.8 (12-step) 0.0 0.0 actionlrlr R L R F F F R F F FFF (a) After 1000 trials from the position (III) 0.0 L FF F (b) After 4000 trials from the position (III) F L F F L F F F L F F (c) After 4000 trials from the position (I) Fig. 6. Some samples of Q-value change in one trial. The weight sets after 1000 trials are used in the case of (a), and that after 4000 trials are used in the case of (b) and (c). The starting location is different between (b) and (c), and refer Fig. 5 for it. would increase as the ideal Q-value that is the exponential curve decided from the discount factor γ and the final reward 0.8. It can be seen by comparing Fig. 6(a) and (b) that in the case of after 1000 trials, the episode length is larger than the case of after 4000 trials. The series of actions chosen can be seen at the bottom of each figure. The black AIBO chose the turn left action even in the case (b) of the front start after 4000 trials. The reason for this is that the black AIBO is likely to turn right slightly even when it chooses the action go forward, and so compensated the shift. Furthermore, in the case (a) of after 1000 trials, the Q values are very different from the ideal, but in the cases (b) and (c) of after 4000 trials, the maximum Q-value is very similar to the ideal. From this result, it can be inferred that learning actually progressed even though the success ratio did not seem to change pronouncedly after 1000 trials. Next, the role of the hidden neuron is guessed from the connection weights. Because the number of connections in each lowest hidden neuron is the same as the number of inputs, the weight pattern after a linear transformation can be observed as a color image whose size is 52 40. The weight patterns seem random due to the influence of the initial connection weight that were determined randomly. However, revealing patterns could be found when the change of each weight from the initial value was observed. The linear transformation from each weight value to the corresponding pixel value is as pixel color,i,j = w af ter,color,i,j w before,color,i,j 127 + 128, (2) max color,i,j w af ter,color,i,j w before,i,j where w after,w before indicate the weight after and before learning respectively. ThecolorcanbeR,G,orB,andi, j indicates the row and column number of a pixel in the image respectively. Most of the weight images are very vague, but in some of them, the image of the white AIBO is suggested as in Fig. 7(a-1,2). The upper images of Fig. 7(a-1,2) are similar camera images selected from actually captured ones. The representations look different from those observed in our previous experiment[3]

760 K. Shibata and T. Kawano (a-1) (a-2) (a) Lower hidden layer (b-4) (b-5) (b) Middle hidden layer (b-1) (b-2) (b-3) Fig. 7. Visualization of the hidden neurons role generated from the connection weights change from the initial values. The images for the lowest hidden neurons are generated by Eq. (2), and those for the middle hidden neurons are generated as the weighted sum of the images for the lowest hidden neurons. The four images at the top are actual images that seem relevant to the weight images. in which a clearer AIBO image could be seen. In the experiment, the black AIBO was fixed and only the head was rotated. Furthermore, the head angle could be one of only 9 states. This means that the number of views of the white AIBO is limited, which was enough for each hidden neuron to represent only one of the views of the target AIBO. However, in this experiment, the number of hidden neurons is not sufficient for each neuron to represent only one view because the number of views is far larger than the number of hidden neurons. In order to roughly see the role of each neuron in the middle hidden layer, the weighted sum of the weight changes of the lowest hidden neurons by the connection weights between the lowest and middle hidden layers is observed. The word roughly is used because the non-linear factor in the lowest hidden layer neurons is neglected. Fig. 7 (b-1,...,5) shows the visualized weight image of some neurons in the middle hidden layer after normalization from 0 to 255. In most of the images, the dark or bright thin band-like area spread laterally in the middle section of the image is clearly visible in Fig. 7(b-1,...,5). It can be inferred that neurons such as (b-1,2,3) contribute to detect the lateral location of the target AIBO. In most of them, by the lateral extension of the band-like area, the color changes from dark to bright or from bright to dark at the same height. It is inferred that this emphasizes the contrast in the neuron s output according to the lateral location of the target AIBO. When comparing with the weight images of (b-4,5), the dark blue area is spread wider to the upper area of the image in (b-4), while the dark area is thin in (b-5). Referring from the actual

Learning of Action Generation from Raw Camera Images 761 camera images above these weight images, it can be inferred that such neurons contribute to detect the distance to the target AIBO. The representations also look different from those observed in our previous experiment in which fat AIBO images often appeared in the images for the neurons in the middle hidden layer. However, the detection of the lateral or forward distance is only a part of the functions that the NN acquired. Neurons in this experiment must acquire a variety of functions such as compensation of light conditions, neglect of background and some other functions that we were not aware of from the connection weights easily. This is very similar to the situation where we cannot understand exactly how the brain functions when we see the excitation pattern of real neurons in the brain. 3 Conclusion It was shown that although our proposed learning system that consists of one neural network trained by reinforcement learning is very simple, it worked on a real walking robot in a real-world-like environment. Many hidden neurons seem to obtain useful feature extraction abilities through learning without any suggestions from the others. The image processing acquired in this task seems different from the template-like image processing that was found in our previous work in which the robot did not walk and the head could rotate to one of only 9 states. This difference shows the flexible and purposive function emergence in a neural network by our proposed approach. Acknowledgments. This research was supported by JSPS Grant in-aid for Scientific Research #19300070. We would like to express our gratitude to Dr. Ito and Mr. Kakiuchi in Naruto University of Education for permitting us to employ their software. References 1. Shibata, K., Okabe, Y.: Reinforcement Learning When the Visual Signals are Directly Given as Inputs. In: Proc. of ICNN 1997, vol. 3, pp. 1716 1720 (1997) 2. Shibata, K., Iida, M.: Acquisition of Box Pushing by Direct-Vision-Based Reinforcement Learning. In: Proc. of SICE Annual Conf. 2003, 0324.pdf, pp. 1378 1383 (2003) 3. Shibata, K., Kawano, T.: Acquisition of Flexible Image Recognition by Coupling of Reinforcement Learning and a Neural Network. The SICE Journal of Control, Measurement, and System Integration 2(2) (to appear, 2009) 4. Sutton, R.S., Barto, A.: Reinforcement Learning: An Introduction, A Bradford Book. MIT Press, Cambridge (1998) 5. Ito, Y., Kakiuchi, H., Oizumi, K., Kikuchi, A.: Development of Educational Software to Control Walking Motion for Four-legged Robot. The Japanese Society of Technology Education 49(1), 1 9 (2007) 6. Watkins, C.J.C.H.: Learning from Delayed Rewards, PhD thesis, Cambridge University, Cambridge, England (1989) 7. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning Internal Representation by Error Propagation. In: Parallel Distributed Processing, vol. 1, pp. 318 364. MIT Press, Cambridge (1986)