Hebbian Plasticity for Improving Perceptual Decisions

Similar documents
Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment. Berkes, Orban, Lengyel, Fiser.

Plasticity of Cerebral Cortex in Development

Experimental Design. Thomas Wolbers Space and Aging Laboratory Centre for Cognitive and Neural Systems

Object recognition and hierarchical computation

A neural circuit model of decision making!

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

Neural Correlates of Human Cognitive Function:

5th Mini-Symposium on Cognition, Decision-making and Social Function: In Memory of Kang Cheng

lateral organization: maps

Visual Categorization: How the Monkey Brain Does It

Perceptual learning as improved probabilistic inference in early sensory areas

Toward the neural causes of human visual perception and behavior

Synaptic theory of gradient learning with empiric inputs. Ila Fiete Kavli Institute for Theoretical Physics. Types of learning

Overview of the visual cortex. Ventral pathway. Overview of the visual cortex

Introduction to Computational Neuroscience

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J.

Memory, Attention, and Decision-Making

Learning without Training

Single cell tuning curves vs population response. Encoding: Summary. Overview of the visual cortex. Overview of the visual cortex

Introduction to Computational Neuroscience

Cell Responses in V4 Sparse Distributed Representation

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014

Representational Switching by Dynamical Reorganization of Attractor Structure in a Network Model of the Prefrontal Cortex

Oxford Foundation for Theoretical Neuroscience and Artificial Intelligence

Mechanisms of generalization perceptual learning

Chapter 5. Summary and Conclusions! 131

Computational Cognitive Neuroscience (CCN)

Carnegie Mellon University Annual Progress Report: 2011 Formula Grant

Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence

Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6)

Observational Learning Based on Models of Overlapping Pathways

Course Introduction. Neural Information Processing: Introduction. Notes. Administration

Computational Cognitive Neuroscience (CCN)

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

fmri: What Does It Measure?

Neural Information Processing: Introduction

Introduction to Computational Neuroscience

Learning in neural networks

Topological Considerations of Memory Structure

Ch.20 Dynamic Cue Combination in Distributional Population Code Networks. Ka Yeon Kim Biopsychology

Prediction of Successful Memory Encoding from fmri Data

Learning and Adaptive Behavior, Part II

Classification and Statistical Analysis of Auditory FMRI Data Using Linear Discriminative Analysis and Quadratic Discriminative Analysis

Dynamic 3D Clustering of Spatio-Temporal Brain Data in the NeuCube Spiking Neural Network Architecture on a Case Study of fmri Data

Bundles of Synergy A Dynamical View of Mental Function

EDGE DETECTION. Edge Detectors. ICS 280: Visual Perception

MSc Neuroimaging for Clinical & Cognitive Neuroscience

Sensory Adaptation within a Bayesian Framework for Perception

Computational Explorations in Cognitive Neuroscience Chapter 7: Large-Scale Brain Area Functional Organization

Choosing the Greater of Two Goods: Neural Currencies for Valuation and Decision Making

Attention Response Functions: Characterizing Brain Areas Using fmri Activation during Parametric Variations of Attentional Load

Sum of Neurally Distinct Stimulus- and Task-Related Components.

Virtual Brain Reading: A Connectionist Approach to Understanding fmri

M Cells. Why parallel pathways? P Cells. Where from the retina? Cortical visual processing. Announcements. Main visual pathway from retina to V1

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science

Sparse Coding in Sparse Winner Networks

acquisition associative learning behaviorism A type of learning in which one learns to link two or more stimuli and anticipate events

On optimal decision-making in brains and social insect colonies Marshall, Bogacz, Dornhaus, Planque, Kovacs,Franks

Dynamics of Color Category Formation and Boundaries

A model to explain the emergence of reward expectancy neurons using reinforcement learning and neural network $

Lateral Geniculate Nucleus (LGN)

Identify these objects

Lecture 2.1 What is Perception?

1. A type of learning in which behavior is strengthened if followed by a reinforcer or diminished if followed by a punisher.

Katsunari Shibata and Tomohiko Kawano

Supplemental Information: Adaptation can explain evidence for encoding of probabilistic. information in macaque inferior temporal cortex

Lesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS

Orientation Specific Effects of Automatic Access to Categorical Information in Biological Motion Perception

The Eye. Cognitive Neuroscience of Language. Today s goals. 5 From eye to brain. Today s reading

OPTO 5320 VISION SCIENCE I

MECHANISMS OF PERCEPTUAL LEARNING

Category Learning in the Brain

A THEORY OF MCCOLLOUGH EFFECT AND. CHUN CHIANG Institute of Physics, Academia Sinica

Supporting Information

Multimodal interactions: visual-auditory

Experimental design for Cognitive fmri

Just One View: Invariances in Inferotemporal Cell Tuning

Introduction. Visual Perception Aditi Majumder, UCI. Perception is taken for granted!

Vision Research 61 (2012) Contents lists available at SciVerse ScienceDirect. Vision Research. journal homepage:

Modeling of Hippocampal Behavior

Frank Tong. Department of Psychology Green Hall Princeton University Princeton, NJ 08544

Memory Processes in Perceptual Decision Making

Basics of Perception and Sensory Processing

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14

Visual Motion Detection

Neural correlates of short-term perceptual learning in orientation discrimination indexed by event-related potentials

A Memory Model for Decision Processes in Pigeons

Biological Information Neuroscience

AU B. Sc.(Hon's) (Fifth Semester) Esamination, Introduction to Artificial Neural Networks-IV. Paper : -PCSC-504

TMS Disruption of Time Encoding in Human Primary Visual Cortex Molly Bryan Beauchamp Lab

Models of Attention. Models of Attention

Visual Context Dan O Shea Prof. Fei Fei Li, COS 598B

Detection of Cognitive States from fmri data using Machine Learning Techniques

Rule-Based Learning Explains Visual Perceptual Learning and Its Specificity and Transfer

How do individuals with congenital blindness form a conscious representation of a world they have never seen? brain. deprived of sight?

A Neural Network Model of Naive Preference and Filial Imprinting in the Domestic Chick

Carl Wernicke s Contribution to Theories of Conceptual Representation in the Cerebral Cortex. Nicole Gage and Gregory Hickok Irvine, California

A model of parallel time estimation

Transcription:

Hebbian Plasticity for Improving Perceptual Decisions Tsung-Ren Huang Department of Psychology, National Taiwan University trhuang@ntu.edu.tw Abstract Shibata et al. reported that humans could learn to repeatedly evoke a stimulus-associated functional magnetic resonance imaging (fmri) activity pattern in visual areas V1/V2 through which visual perceptual learning was achieved without stimulus presentation. Contrary to their attribution of visual improvements to neuroplasticity in adult V1/V2, our Hebbian learning interpretation of these data explains the attainment of better perceptual decisions without plastic V1/V2. Keywords: Categorization, Decision Making, Multi-Voxel Pattern Analysis, Neurofeedback, Neuroplasticity, Perceptual Learning, Reinforcement Learning

Through on-line feedback, Shibata et al. (1) demonstrated the ability of humans to induce a stimulus-associated functional magnetic resonance imaging (fmri) voxel pattern in the early visual areas V1/V2. Furthermore, visual sensitivity to an oriented grating stimulus in a noisy background was improved through repeated induction of such a voxel pattern. Interestingly, no grating-related perceptual experience was evoked in their participants by this perceptual learning paradigm. How can stimulus-specific perceptual improvement be attained without external presentation and internal imagery of the stimulus? What exactly has been learned by the neurofeedback participants? Category Units Hebbian Learning Voxel Pa ern Reinforcement Learning Sub-voxel Ac vi es (Cogni ve Strategies) Fig. 1. Two proposed types of learning involved in the neurofeedback paradigm. In this schematic, all voxel cells are connected with all category units, and their relative strength of connection weights are learned through Hebbian mechanisms. For reinforcement learning, different cognitive strategies and sub-voxel activities may activate the same pattern at the voxel level. 1

Our explanation for these data is that both reinforcement and unsupervised Hebbian learning are involved in their neurofeedback procedure (Fig. 1). The reinforcement signal was a variable-size disk indicating the degree of match between a spontaneous voxel pattern in a trial and a stimulus-associated template pattern. The induced template-like patterns in V1/V2 then allowed unsupervised Hebbian mechanisms to strengthen the synaptic connections between repetitively activated V1/V2 neurons and their receiving neurons, such as those in V3. In effect, exposure-specific perceptual decisions were improved because neurons in downstream areas learned to respond more to regular grating signals and to ignore irregular background noises in the inputs. Owing to the limited spatial resolution of fmri, the possibility remains that different cognitive strategies and neurons at the sub-voxel level could activate the same voxel pattern. Therefore, induction of a grating-associated voxel pattern did not necessarily trigger perception of that grating. This interpretation, contrary to the conclusion of Shibata et al., attributes visual improvements in the neurofeedback study to neuroplasticity in brain structures other than V1/V2. Note that their stimulus-associated template pattern was a voxel pattern obtained before perceptual learning, and their reinforcement procedure in fact encouraged preservation of such a pre-training stimulus representation in V1/V2. Because unchanged neural representations cannot account for changed behavioral performances, the observed improvement in perceptual decisions likely reflects neuroplasticity somewhere else in the visual system, such as areas V3 and V4, whose response properties were free to change during the V1/V2 pattern induction stage. In general, visual perceptual learning (VPL) can result from changes of neuronal responses in early visual areas V1/V2 (2-4) and/or higher brain areas that underlie visual recognition, attention and decision making (5,6). Although the neurofeedback results support for the latter case from the viewpoint of Hebbian learning, our suggested Hebbian plasticity between connecting neurons may, in principle, occur within or across any brain areas, including V1/V2, in other VPL tasks. Moreover, changes of neural connectivity and response in a neuronal network are often two faces of the same coin neither characterizes VPL. For example, changed connectivity between the lateral geniculate nucleus and V1 alters V1 neuronal response in an orientation discrimination task (4), and 2

changed connectivity between the middle temporal and lateral intraparietal (LIP) areas alters LIP neurononal response in a motion discrimination task (6). To illustrate Hebbian plasticity for VPL, pattern-specific improvement in perceptual decisions was simulated by a two-layer neural network with divisive/shunting normalization in each layer (7). The network matches an input pattern x! with each category prototype w! j and outputs the degree of match y j to category units:!! y = w x. j The best match category J is chosen as the perceptual decision, and its category prototype then updates to accommodate the current instance based on an unsupervised Hebbian learning algorithm known as the Instar rule (8), or conditional principal component analysis (9): Δ w J ~ y J ( x w J ). j 100 Pattern 1 (N=20) 100 Pattern 2 (N=20) 100 Pattern 3 (N=20) 90 90 90 80 80 80 Accuracy (%) 70 60 50 Pre test Post test Accuracy (%) 70 60 50 Pre test Post test Accuracy (%) 70 60 50 Pre test Post test 40 40 40 0.2 0.5 1.5 10 S/N Ratio 0.2 0.5 1.5 10 S/N Ratio 0.2 0.5 1.5 10 S/N Ratio Fig. 2. Simulation results of improved perceptual decisions after unsupervised Hebbian learning [cf. Fig. 3A in (1)]. Blue and red points indicate classification performances of a 3x3 input pattern with different signal-to-noise (S/N) ratios before and after Hebbian learning. Classification accuracy only improves for a repeatedly presented signal pattern, as shown in the middle panel. Error bars depict the standard errors of the group means from 20 independent simulations with randomized initial weights for network connectivity and randomized presentation order of stimuli to the network. 3

After learning, a category prototype is approximately equal to the weighted average of category instances, with weights being the degree of match. In other words, the network incrementally learns the regularities in the input patterns, particularly from the low-noise instances. The simulation results (Fig. 2) for a three-alternative forcedchoice task qualitatively resemble the behavioral performance reported by Shibata et al. in that decision accuracy is a monotonic function of signal-to-noise ratio, and it improves only for the choice corresponding to a repeatedly presented signal pattern. Note that such a Hebbian model can also account for VPL with feedback by adding supervised decision signals into the category units. Thus, it is a general model for learning in perceptual decision making with multiple choices and goes beyond VPL models for binary decisions [e.g., (4, 10-11)]. For the neurofeedback study, whereas the perceptual learning component is explicable, the reinforcement-learning (RL) problem faced by participants is challenging in theory due to its infinite state and action spaces. Unlike conventional RL problems with delayed rewards, the neurofeedback study provided immediate evaluative feedback in each trial. The challenge for the participants was to discover effective actions by trial and error before reward-bearing actions could be further exploited. On balance, sufficient exploration is the key to success in reinforcement learning (12). Because participants had no access to the state/voxel space, to optimize reward they had to estimate reward as a function of an experimentally unrestricted, infinite set of action options. In reality, the participants might randomly or parametrically explore a finite set of N actions. Because the overall reward function and its derivatives were unknown with respect to the N actions under exploration, this optimization process required sampling of the N-dimensional continuous-valued action space and was expected to be slow with a large N (13). Given that action evaluation was sequential in the neurofeedback study and N might dynamically increase when the chosen actions were reward-irrelevant, it is unclear how participants could quickly learn to induce V1/V2 activity patterns within 180 attempts (i.e., the total trial number of a pattern induction session). In summary, we offer a theoretical perspective on the neurofeedback study by Shibata et al., who substituted stimulus-derived biofeedback signals for real stimuli in 4

visual perceptual training. The lack of stimulus perception might result from other stimulus-unrelated sub-voxel activities that shared the same template pattern at the voxel level. Together with the procedure of clamping a pre-training pattern, it also suggests the observed improvement of perceptual decisions to be a refinement of inference about the input signals rather than the stimulus representation in V1/V2 per se. Such an inference improvement may be attributed to synaptic plasticity in Hebbian learning, which is a viable mechanism for other forms of perceptual learning (10, 14). Finally, it remains an open question why rapid learning of pattern induction was observed in the neurofeedback study given that uninformed, sequential exploration of an infinite state/action space is theoretically a slow process. 5

References and Notes 1. K. Shibata, T. Watanabe, Y. Sasaki, M. Kawato, Perceptual learning incepted by decoded fmri neurofeedback without stimulus Presentation. Science 334, 1413-1415 (2011). 2. A. Karni, D. Sagi, The time course of learning a visual skill. Nature 365, 250-252 (1993). 3. A. A. Schoups, R. Vogels, N. Qian, G. A. Orban, Practising orientation identification improves orientation coding in V1 neurons. Nature 412, 549-553 (2001). 4. V. R. Bejjanki, J. M. Beck, Z.-L. Lu, A. Pouget, Perceptual learning as improved probabilistic inference in early sensory areas. Nat. Neurosci. 14, 642-648 (2011). 5. T. Yang, J. H. R. Maunsell, The effect of perceptual learning on neuronal responses in monkey visual area V4. J. Neurosci. 24, 1617-1626 (2004). 6. C.-T. Law, J. I. Gold, Neural correlates of perceptual learning in a sensory-motor, but not a sensory, cortical area. Nat. Neurosci. 11, 505-513 (2008). 7. S. Grossberg, Contour enhancement, short-term memory, and constancies in reverberating neural networks. Stud. Appl. Math. 52, 213-257 (1973). 8. S. Grossberg, Adaptive pattern classification and universal recoding, I: Parallel development and coding of neural feature detectors. Biol. Cybern. 23, 121-134 (1976). 9. R. C. O'Reilly, Y. Munakata, Computational Explorations in Cognitive Neuroscience: Understanding the Mind by Simulating the Brain (MIT Press, Cambridge, MA, 2000). 10. A. Petrov, B. Dosher, Z.-L. Lu, The dynamics of perceptual learning: An incremental reweighting model. Psychol. Rev. 112, 715-743 (2005). 11. C.-T. Law, J. I. Gold, Reinforcement learning can account for associative and perceptual learning on a visual-decision task. Nat. Neurosci. 12, 655-663 (2009). 12. R. S. Sutton, A. G. Barto, Reinforcement Learning: An Introduction (MIT Press, Cambridge, MA, 1998). 13. T. G. Kolda, R. M. Lewis, V. Torczon, Optimization by direct search: New perspectives on some classical and modern methods. SIAM Rev. 45, 385-482 (2003). 14. L. M. Saksida, J. L. McClelland, in The Cognitive Neuroscience of Memory Encoding and Retrieval, A. Parker, T. Bussey, E. Wilding, Eds. (Psychology Press, Hove, UK, 2002), pp. 259-282. 6