Modulators of Spike Timing-Dependent Plasticity

Similar documents
Shadowing and Blocking as Learning Interference Models

Basics of Computational Neuroscience: Neurons and Synapses to Networks

Beyond Vanilla LTP. Spike-timing-dependent-plasticity or STDP

Recognition of English Characters Using Spiking Neural Networks

VS : Systemische Physiologie - Animalische Physiologie für Bioinformatiker. Neuronenmodelle III. Modelle synaptischer Kurz- und Langzeitplastizität

Evaluating the Effect of Spiking Network Parameters on Polychronization

Temporally asymmetric Hebbian learning and neuronal response variability

Neuromorphic computing

Synaptic Transmission: Ionic and Metabotropic

Modeling Depolarization Induced Suppression of Inhibition in Pyramidal Neurons

Computational cognitive neuroscience: 2. Neuron. Lubica Beňušková Centre for Cognitive Science, FMFI Comenius University in Bratislava

Part 11: Mechanisms of Learning

Efficient Emulation of Large-Scale Neuronal Networks

Spiking neural network simulator: User s Guide

A general error-based spike-timing dependent learning rule for the Neural Engineering Framework

Realization of Visual Representation Task on a Humanoid Robot

Multi compartment model of synaptic plasticity

Investigation of Physiological Mechanism For Linking Field Synapses

International Journal of Advanced Computer Technology (IJACT)

Solving the distal reward problem with rare correlations

Heterogeneous networks of spiking neurons: self-sustained activity and excitability

Lab 4: Compartmental Model of Binaural Coincidence Detector Neurons

Implementing Spike-Timing-Dependent Plasticity on SpiNNaker Neuromorphic Hardware

Stable Hebbian Learning from Spike Timing-Dependent Plasticity

Intro. Comp. NeuroSci. Ch. 9 October 4, The threshold and channel memory

Tunable Low Energy, Compact and High Performance Neuromorphic Circuit for Spike-Based Synaptic Plasticity

Computational Approaches in Cognitive Neuroscience

Part 3. Synaptic models

Synchrony-Induced Switching Behavior of Spike Pattern Attractors Created by Spike-Timing-Dependent Plasticity

Signal detection in networks of spiking neurons with dynamical synapses

Parametrization of Neuromodulation in Reinforcement Learning

Modeling synaptic facilitation and depression in thalamocortical relay cells

Synaptic Plasticity and the NMDA Receptor

Solving the distal reward problem with rare correlations

STDP between a Pair of Recurrently Connected Neurons with Asynchronous and Synchronous Stimulation AZADEH HASSANNEJAD NAZIR

Temporally Asymmetric Hebbian Learning, Spike Timing and Neuronal Response Variability

Introduction to Computational Neuroscience

Adaptive leaky integrator models of cerebellar Purkinje cells can learn the clustering of temporal patterns

Observational Learning Based on Models of Overlapping Pathways

Active Control of Spike-Timing Dependent Synaptic Plasticity in an Electrosensory System

Dynamic Stochastic Synapses as Computational Units

Self-Organization and Segmentation with Laterally Connected Spiking Neurons

Neuroscience Flythrough. Lukas Schott und Letita Parcalabescu

Why do we have a hippocampus? Short-term memory and consolidation

A Model of Spike-Timing Dependent Plasticity: One or Two Coincidence Detectors?

Neuron Phase Response

Synaptic Plasticity and Connectivity Requirements to Produce Stimulus-Pair Specific Responses in Recurrent Networks of Spiking Neurons

Consciousness as representation formation from a neural Darwinian perspective *

Long-term synaptic plasticity. N500, 6 Sept., 2016

Theme 2: Cellular mechanisms in the Cochlear Nucleus

arxiv: v1 [q-bio.nc] 25 Apr 2017

A SUPERVISED LEARNING

Rolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.

Simulating inputs of parvalbumin inhibitory interneurons onto excitatory pyramidal cells in piriform cortex

LEARNING ARBITRARY FUNCTIONS WITH SPIKE-TIMING DEPENDENT PLASTICITY LEARNING RULE

Brain Rhythms and Mathematics

How has Computational Neuroscience been useful? Virginia R. de Sa Department of Cognitive Science UCSD

Relative contributions of cortical and thalamic feedforward inputs to V2

You submitted this quiz on Sun 19 May :32 PM IST (UTC +0530). You got a score of out of

Hebbian Learning and Plasticity

Preparing More Effective Liquid State Machines Using Hebbian Learning

Indirect Training of a Spiking Neural Network for Flight Control via Spike-Timing-Dependent Synaptic Plasticity

Reinforcement Learning in a Neurally Controlled Robot Using Dopamine Modulated STDP

Modeling of Hippocampal Behavior

Technischer Bericht TUM. Institut für Informatik. Technische Universität München

COMPLEX NEURAL COMPUTATION WITH SIMPLE DIGITAL NEURONS. Andrew Thomas Nere. A dissertation submitted in partial fulfillment of

Hearing Thinking. Keywords: Cortical Neurons, Granular Synthesis, Synaptic Plasticity

Computational analysis of epilepsy-associated genetic mutations

Spiking Inputs to a Winner-take-all Network

Synaptic plasticityhippocampus. Neur 8790 Topics in Neuroscience: Neuroplasticity. Outline. Synaptic plasticity hypothesis

Introduction to Computational Neuroscience

Different inhibitory effects by dopaminergic modulation and global suppression of activity

Spike-Timing Dependent Plasticity, Learning Rules

Dynamics of Hodgkin and Huxley Model with Conductance based Synaptic Input

Synapses and synaptic plasticity. Lubica Benuskova Lecture 8 How neurons communicate How do we learn and remember

The Role of Coincidence-Detector Neurons in the Reliability and Precision of Subthreshold Signal Detection in Noise

Cellular Neurobiology BIPN140

Imperfect Synapses in Artificial Spiking Neural Networks

Temporal asymmetry in spike timing-dependent synaptic plasticity

Cell Responses in V4 Sparse Distributed Representation

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

STDP enhances synchrony in feedforward network

Supervised Learning in Spiking Neural Networks with ReSuMe: Sequence Learning, Classification and Spike-Shifting

Theory of correlation transfer and correlation structure in recurrent networks

Hiroki Mori Graduate School of Fundamental Science and Engineering Waseda University Tokyo, Japan

Learning in neural networks by reinforcement of irregular spiking

Task 1: Machine Learning with Spike-Timing-Dependent Plasticity (STDP)

Effects of Inhibitory Synaptic Current Parameters on Thalamocortical Oscillations

Model-Based Reinforcement Learning by Pyramidal Neurons: Robustness of the Learning Rule

Lecture 1: Neurons. Lecture 2: Coding with spikes. To gain a basic understanding of spike based neural codes

QUIZ/TEST REVIEW NOTES SECTION 7 NEUROPHYSIOLOGY [THE SYNAPSE AND PHARMACOLOGY]

Running PyNN Simulations on SpiNNaker

What is Anatomy and Physiology?

Temporal coding in the sub-millisecond range: Model of barn owl auditory pathway

Designing Behaviour in Bio-inspired Robots Using Associative Topologies of Spiking-Neural-Networks

BIONB/BME/ECE 4910 Neuronal Simulation Assignments 1, Spring 2013

Self-organizing continuous attractor networks and path integration: one-dimensional models of head direction cells

SUPPLEMENTARY INFORMATION

The synaptic Basis for Learning and Memory: a Theoretical approach

Neurobiology: The nerve cell. Principle and task To use a nerve function model to study the following aspects of a nerve cell:

Transcription:

Modulators of Spike Timing-Dependent Plasticity 1 2 3 4 5 Francisco Madamba Department of Biology University of California, San Diego La Jolla, California fmadamba@ucsd.edu 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 Abstract Spike Timing-Dependent Plasticity (STDP) is a Hebbian form of learning that captures the temporal relationship between pre- and postsynaptic spikes. Recent studies have uncovered that the relative concentration of neuromodulators, in addition to the timing of pre- and postsynaptic spikes, can affect the polarity of STDP. In this paper we sought to modify existing STDP rule implementing neural networks to account for experimental findings and observe the corresponding changes in the synaptic weights. Initial conclusions support the notion that minimal changes in A P or A D have little effect on rate of synaptic changes however further work must be done to verify or refute these findings. 1 Introduction Long-term potentiation (LTP) and long-term depression (LTD) are well known mechanisms for cortical synaptic plasticity. Traditionally, the given outcome, LTP or LTD, was determined by varying the presynaptic firing rate or the postsynaptic membrane potential during conditioning. With the discovery that when a presynaptic spike precedes a post synaptic spike by several seconds LTP results and when a postsyanptic spike precedes a postsynaptic spike by several seconds LTD results, the temporal relationship between the presynaptic and postsynaptic spike became an important factor in the study of synaptic plasticity. The discovery of spike timing-dependent plasticity [1] [2] has led to many developments in the study of neural networks. It was found that oscillations greatly facilitate STDP's ability to detect repeating patterns. Specifically, Masquelier et al.[3] were able to demonstrate, in simulations, how a single neuron equipped with STDP robustly detects a pattern of input currents automatically encoded in the phases of a subset of its afferent s, and repeating at random intervals. In addition it was demonstrated that a model network of cortical spiking neurons modulated with the neuromodulator dopamine (DA) could solve the distal reward problem [4]. Surprisingly, much of the recent progress of STDP involves implementing the STDP rule into neural networks, and not on the STDP rule. Only recently has significant progress been made to control the rules of STDP, specifically the given outcome and the time window for coincidence detection. A recent study found that the relative activation of neuromodulator receptors coupled to adenylyl cyclase (AC) and phospholipase C (PLC) signaling cascades can override the order of the timing of the pre- and postsynaptic spikes [5]. In addition, they found that these modulators could also increase the window of coincidence detection and effectively "prime" neuron synapses into plastic states.

43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 The aim of this proect was to attempt to modify existing STDP rule implementing neural networks to account for experimental findings on neuromodulators. Since the experimental findings were found in an in vitro system, and hence their release and uptake were tightly controlled, biologically relevant concentration and kinetics could not be obtained. As a result, the focus of this proect was to model the possible changes that addition of particular neuromodulators can have on the system. 2 Methods 2.2 Leaky Integrate and Fire Neurons In single neuron or simple neuronal networks, conductance-based models, such as the Hodgkin-Huxley (1952) and Morris-Lecar (1981) models, can be useful because such models contains variables and parameters that have well-defined biophysical meanings that can be measured experimentally [6]. However when neuronal network contain many constituents, the number of connections and the vast amount of differential equations that must be solved can become computationally taxing. Furthermore, such models, which contain detailed accounts of axon potential generation and conductance, often obtain parameters from many neurons which are averaged to produce a value. For our present purposes, it is thus convenient to use a simple model which faithfully reproduces the most important neurcomputational features of the neuron (voltage threshold, reset, membrane potential) and omit some of the more taxing features (detailed action potential generation and propagation). Leaky Integrate-and-fire neurons (LIF neuron) are the simplest possible neural model. In this model, complicated conductances responsible for action potential generation have been omitted and instead the spiking mechanism is replaced by a potential threshold, v thres [7]. When v(t 1 ) = v thres has been reached, a spike it emitted at t 1 and the potential is reset to zero. In this model, the membrane potential follows the differential equation: C dv dt = v + I, t > 0 R where I = I(t) is a stimulation current and v(0) = 0 is an initial condition. Without input current and at steady state, the membrane potential is equal to the resting membrane potential value of the model, which in this case is zero. If an input current is present, the synaptic inputs to a LIF neuron can be simulated as a conductance change: n ex I syn (t) = g ex (t, t ex,n )(v v ex ) + g in (t t in,n )(v v in ) n=1 Due to their relative simplicity, LIF neurons can be easily expanded to LIF networks. LIF neurons in networks operate according to a specific order: (1) check for an input spike at the current time, t, and for networks spikes from previous time, t-dt; (2) update conductances based on the input spikes and network spikes recorded in (1); (3) update potentials, recorded spikes, and return to (1). Finally, LIF networks can be modeled with the STDP rule resulting in plastic network. The network equations for the weight increases and decreases are presented below which occur when the LIF network order (see above) reports that cell k fires in the interval [dt, (+1)dt). In order to limit the maximum weight gain, the potentiation of the presynaptic weight is hard bounded by (W max W k,kpre ) and the depression of the postsynaptic weights n in n=1

93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 is hard bounded by W kpost,k. +1 W k,kpre = W k,kpre +1 W kpost,k = W kpost,k + A P exp ( T k pre ( + 1)dt ) (W τ max W k,kpre ) P 2.2 Neuromodulator sources and kinetics + A D exp ( T k post ( + 1)dt ) W τ kpost,k D In their paper, Seol [5] identified isoproterenol (Iso) and 4-[N-(3-Chlorophenyl) carbamoyloxy]-2-butynyltrimethylammonium chloride (McN) as the neuromodulators that (McN) and that induced associative LTP or LTD respectively. Iso is an agonist of β- adrenergic receptors coupled to the AC cascade and Mcn is a muscarinic cholinergic receptor M1 agonist coupled to the PLC cascade. In order to model their effects, the equations used by Izhikevich (2007) [4] to model the effects of dopamine were adapted to account for Iso: P AC = P AC + STDP(t τ post t pre )δ(t t pre ) PAC A Iso S P = P AC A Iso = B Iso τ Iso + ISO(t) Where P AC refers to the activation of an enzyme important for plasticity; A ISO describes the extracellular concentration of Iso; δ(t t pre ) is the Dirac delta function that step-increases post the variable P AC ; S P is the potentiating contribution to the synaptic weight; τ ISO refers to the time constant of Iso uptake or its diffusion away from the receptors; and ISO(t) models the source of Iso. Likewise, the equations for the effects of McN are below with parameters exactly as for Iso but adapted for McN. D PLC post = D PLC + STDP(t τ post t pre )δ(t t pre ) DPLC B McN 2.2 Weight increase and decreases s D = D PLC B McN = B McN τ McN + McN(t) The experiments described by Seol [5]were conducted in an in vitro system consisting of layer IV to layer II/III inputs in the rat visual cortex. Although they modulated STDP, Iso and McN represent foreign neuromodulators and hence the values for their concentration at the synapse, their uptake rates, and production rates have not yet been (or simply cannot be) obtained. The values for dopamine found in Izhikevich's paper could not be used due to fundamentally different situations. In Izhikevich's paper [4], dopamine served as an extracellular reward used to identify the correct source of STDP and reinforce the correct pre- post synaptic connection. post

135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 In our present paper, neuromodulators directly increase or decrease the weights at the synapse and do not serve to pick out a particular synaptic connection. As a result, until biological relevant neuromodulators are identified in vivo, as well as their kinetics measured, we will proceed in our experiment to absorb the values of S P into A P and the values of s D into A D. +1 W k,kpre 3 Results = W k,kpre +1 W kpost,k = W kpost,k 3.1 Establishment of controls + A P exp ( T k pre ( + 1)dt ) (W τ max W k,kpre ) P + A D exp ( T k post ( + 1)dt ) W τ kpost,k D A neural network (with the STDP rule) [8] of 100 neurons, 80 excitatory cells and 20 inhibitory cells, was used to establish a baseline value (the code is found in [8]). The matrix of weights between excitatory-excitatory, excitatory-inhibitory, and inhibitoryexcitatory cells was set at 25% connectivity whereas the matrix of weights of inhibitory - inhibitory cells was set at 5%. A simultaneous input was delivered to the excitatory conductance s of the first 16 excitatory cells causing them to fire synchronously ust before the simulation. A simulation with A P = 0.1, A D = 0.3, with a time course of two seconds was run (Fig 1). A plot of synaptic weights at the beginning of the simulation (Fig 1A, 1C) and after (Fig 1B, 1D) clearly show that following the simulation, the synaptic weights of the first 16 excitatory neurons show decreased strength, evident in the brighter plots found in Fig 1B, 1D compared to Fig 1A, 1C. As the first 16 neurons were simultaneously driven by an input stimulus, the synaptic connections between neighboring neurons were expected to degrade over time as the external input would override the minimal contribution of the synaptic weights. The STDP rule in the neural network also causes synaptic weights to either increase to a maximal value (in this case, 0.2) or decrease to zero over time (Fig 1E). The value of A D = 0.3 is represented by the top blue dots; the value of A P = 0.1 is represented by the bottom blue dots. The greater value for A D compared to A P results in synaptic weights decreasing much quicker to 0 then increasing to 0.2 Fig 1F illustrates the initial random distribution of possible synaptic weights before the simulation has begun (red trace) and after the simulation has completed (green trace); as time passes, the fraction of neurons at synaptic values of 0 or 0.2 increases.

A B C D E

174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 F Figure 1: Baseline values. (A) A plot of synaptic weights before the simulation; A P = 0.1, A D = 0.3, total time = 2000msec; x and y axis denote neuron number; red lines differentiate the weight matrices; the red circle encloses the first 16 excitatory cells (B) A plot of synaptic weights after the simulation. (C) Blow up the 16 excitatory cells in (A) before the simulation. (D) Blow up of the 16 excitatory cells in (B) following the simulation. (E) Plot of the fraction of synaptic weights over time (in msec); top blue dots represents fraction of synaptic weights at 0; bottom blue dots represents fraction of synaptic weights at 0.2 (F) The red trace is distribution of synaptic weights at the beginning of the simulation. The green trace is the distribution of synaptic weights at the end of the simulation. 3.2 Neuromodulators and increase in synaptic weights In order to replicate the effects of Iso neuromodulator addition, the values for A P were increased in steps of 0.1 while the value of A D remained constant at 0.3 (Fig 2A-D). Surprisingly all changes in A P resulted in identical plots which would either suggest that small variations in A P would have no effect on the rate of synaptic weight changes or the code is unresponsive to the changes. A

B C 191 Figure 2: Variations in A P ; unless noted, top blue dots represent fraction of synaptic weights D

192 193 194 195 196 197 198 199 200 201 at 0 (A) A P = 0.1, A D = 0.3, total time = 2000msec (B) A P = 0.2, A D = 0.3, total time = 2000msec (C) A P = 0.3, A D = 0.3, total time = 2000msec (D) A P = 0.4, A D = 0.3, total time = 2000msec 3.3Neuromodulators that decrease in synaptic weights In order to replicate the effects of Iso neuromodulator addition, the values for A D were increased in steps of 0.1 while the value of A P remained constant at 0.1 (Fig 3A-D). Interestingly the results found in changing A P while holding A D fixed appeared identical to the ones obtained by changing A D while holding A P fixed. A B

C D 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 Figure 3: Variations in A D (A) A P = 0.1, A D = 0.3, total time = 2000msec (B) A P = 0.1, A D = 0.4, total time = 2000msec (C) A P = 0.1, A D = 0.1, total time = 2000msec (D) A P = 0.1, A D = 0.5, total time = 2000msec 4 Conclusion The similarity in the results found in modifying A P while holding A D fixed and reverse situation would lead to two different conclusions: either the network it insensitive to minimal changes of A P or A D or simply the code was hardcoded to account for only a particular value of A P or A D. Although the results would agree with the latter, extensive review of the code would lead to no indication that the values of A P or A D are hard coded in the system. Recent studies also support the conclusion that neural networks supporting the STDP rule are incredibly efficient at learning. Masquelier et al.[3] have found that in the context of learning with phase-of-firing coding (PoFC); learning was possibly even when only ~10% of afferents exhibited PoFC. Further testing on the code by the owners[8] have shown that the STDP rule applied to the random networks provides it the ability of pattern completion. Ultimately, additional testing in other models [9] is needed to determine whether STDP neural networks are insensitive to small changes in A P or A D. Also, consultation of the

222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 literature for larger values of A P or A D would be informative though on a first past, no such values were discovered. 5 Acknowledgements Thanks to Jeff Bush for his computational input and feedback as well as support and also to Professor Cauwenberghs for providing ideas on how to approach the topic. In addition, thanks to Fabrizio Gabbiani and Steven J. Cox for writing a terrific text on neuroscience and providing an accompanying Matlab code. References [1] Bi, G.Q., Poo, M.M (1998) Synaptic modifications in cultured hippocampal neurons: dependence on spike timing, synaptic strength, and postsynaptic cell type Journal of Neuroscience 18(24):10464-72 [2] Bi, G. & Poo, M. (2001) Synaptic modification of correlated activity: Hebb's postulate revisited. Ann. Rev. Neurosci. 24:139-166 [3] Masquelier, T., Hugues, E., Deco, G. & Thorpe, S.J. (2009) Oscillations, phase-of-firing coding, and spike timing-dependent plasticity: an efficient learning scheme. Journal of Neuroscience 29(43):13484-93. [4] Izhikevich, E.M. (2007) Solving the Distal Reward Problem through Linkage of STDP and Dopamine Signaling. Cerebral Cortex 17(10):2443-2452. [5] Seol, G.H., & Ziburkus, J. (2007) Neuromodulators Control the Polarity of Spike-Timing- Dependent Synaptic Plasticity Neuron 55(6):919-929 [6] Izhikevich, E.M. (2007) Dynamical Systems In Neuroscience. Massachusetts: MIT Press books. [7] Gerstner, W., Kistler, W. (2002) Spiking Neuron Models. Cambridge: Cambridge University Press [8] Gabbiani, F. & Cox, S.J. (2010) Mathematics for Neuroscientists. London: Academic Pess/Elsevier [9] Legenstein, R., Pecevski, D., Maass, W. (2008) A learning theory for reward-modulated spiketiming-dependent plasticity with application to biofeedback. PLoS Comput Biol 4(10):e1000180