19th AWCBR (Australian Winter Conference on Brain Research), 2001, Queenstown, AU

Similar documents
Rolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

Unit 1 Exploring and Understanding Data

Information Processing During Transient Responses in the Crayfish Visual System

Neuromorphic computing

Part 11: Mechanisms of Learning

A Dynamic Neural Network Model of Sensorimotor Transformations in the Leech

Ideal Observers for Detecting Motion: Correspondence Noise

Cerebral Cortex. Edmund T. Rolls. Principles of Operation. Presubiculum. Subiculum F S D. Neocortex. PHG & Perirhinal. CA1 Fornix CA3 S D

Ultrastructural Contributions to Desensitization at the Cerebellar Mossy Fiber to Granule Cell Synapse

A general error-based spike-timing dependent learning rule for the Neural Engineering Framework

arxiv: v1 [q-bio.nc] 25 Apr 2017

Ideal Observers for Detecting Human Motion: Correspondence Noise.

Learning to Use Episodic Memory

Nature Neuroscience: doi: /nn Supplementary Figure 1. Behavioral training.

Neural Correlates of Human Cognitive Function:

Fig. 4. The activity of Pkc -transduced neurons is required for enhanced learning. After gene transfer, rats were tested on [] vs. +.

Anatomy of a Neuron. Copyright 2000 by BSCS and Videodiscovery, Inc. Permission granted for classroom use. Master 2.1

Dynamic Stochastic Synapses as Computational Units

Supplementary Figure 1. Basic properties of compound EPSPs at

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

The Use of Unidimensional Parameter Estimates of Multidimensional Items in Adaptive Testing

Shadowing and Blocking as Learning Interference Models

Meta-Analysis David Wilson, Ph.D. Upcoming Seminar: October 20-21, 2017, Philadelphia, Pennsylvania

A Race Model of Perceptual Forced Choice Reaction Time

Modeling of Hippocampal Behavior

PRACTICAL STATISTICS FOR MEDICAL RESEARCH

PIB Ch. 18 Sequence Memory for Prediction, Inference, and Behavior. Jeff Hawkins, Dileep George, and Jamie Niemasik Presented by Jiseob Kim

Free recall and recognition in a network model of the hippocampus: simulating effects of scopolamine on human memory function

Auditory Scene Analysis

Computational Explorations in Cognitive Neuroscience Chapter 7: Large-Scale Brain Area Functional Organization

Representational Switching by Dynamical Reorganization of Attractor Structure in a Network Model of the Prefrontal Cortex

A Race Model of Perceptual Forced Choice Reaction Time

Classification of Synapses Using Spatial Protein Data

Beyond Vanilla LTP. Spike-timing-dependent-plasticity or STDP

CHAPTER I From Biological to Artificial Neuron Model

Learning. Learning: Problems. Chapter 6: Learning

Neuroscience 201A (2016) - Problems in Synaptic Physiology

2012 Course: The Statistician Brain: the Bayesian Revolution in Cognitive Sciences

Why do we have a hippocampus? Short-term memory and consolidation

Υπολογιστική διερεύνηση των νευρικών μηχανισμών μνήμης και μάθησης

STRUCTURAL ELEMENTS OF THE NERVOUS SYSTEM

A developmental learning rule for coincidence. tuning in the barn owl auditory system. Wulfram Gerstner, Richard Kempter J.

Machine Learning! Robert Stengel! Robotics and Intelligent Systems MAE 345,! Princeton University, 2017

Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment. Berkes, Orban, Lengyel, Fiser.

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J.

ERA: Architectures for Inference

Physiology of Tactile Sensation

CONCEPT LEARNING WITH DIFFERING SEQUENCES OF INSTANCES

Temporally asymmetric Hebbian learning and neuronal response variability

Neural coding and information theory: Grandmother cells v. distributed codes

Bioscience in the 21st century

Observational Learning Based on Models of Overlapping Pathways

Ameen Alsaras. Ameen Alsaras. Mohd.Khatatbeh

Synaptic Learning Rules and Sparse Coding in a Model Sensory System

A model of parallel time estimation

Synaptic plasticityhippocampus. Neur 8790 Topics in Neuroscience: Neuroplasticity. Outline. Synaptic plasticity hypothesis

Introduction to Computational Neuroscience

2012 Course : The Statistician Brain: the Bayesian Revolution in Cognitive Science

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

Local Image Structures and Optic Flow Estimation

Key words: consciousness; information model; system resource architecture; system design

Effective And Optimal Storage of Memory Patterns With Variable Coding Levels

Salient Feature Coding in Response Selection. Circularity?

Katsunari Shibata and Tomohiko Kawano

COGNITIVE SCIENCE 107A. Hippocampus. Jaime A. Pineda, Ph.D.

Algorithms in Nature. Pruning in neural networks

Learning = an enduring change in behavior, resulting from experience.

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014

You submitted this quiz on Sun 7 Apr :08 PM IST (UTC +0530). You got a score of out of Your Answer Score Explanation

A Memory Model for Decision Processes in Pigeons

Action potential. Definition: an all-or-none change in voltage that propagates itself down the axon

A model of the interaction between mood and memory

Sensory Adaptation within a Bayesian Framework for Perception

Plasticity of Cerebral Cortex in Development

Medical Statistics 1. Basic Concepts Farhad Pishgar. Defining the data. Alive after 6 months?

Computerized Mastery Testing

Memory: Computation, Genetics, Physiology, and Behavior. James L. McClelland Stanford University

Understanding Uncertainty in School League Tables*

Free classification: Element-level and subgroup-level similarity

Chapter 5: Field experimental designs in agriculture

Different inhibitory effects by dopaminergic modulation and global suppression of activity

Natural Scene Statistics and Perception. W.S. Geisler

Automaticity of Number Perception

Synaptic Plasticity and Connectivity Requirements to Produce Stimulus-Pair Specific Responses in Recurrent Networks of Spiking Neurons

Computational Cognitive Neuroscience (CCN)

Sum of Neurally Distinct Stimulus- and Task-Related Components.

Chapter 12: Fundamentals of the Nervous System and Nervous Tissue

Measuring the User Experience

DANIEL KARELL. Soc Stats Reading Group. Princeton University

26:010:557 / 26:620:557 Social Science Research Methods

Reading Neuronal Synchrony with Depressing Synapses

Chapter 5: Learning and Behavior Learning How Learning is Studied Ivan Pavlov Edward Thorndike eliciting stimulus emitted

Synfire chains with conductance-based neurons: internal timing and coordination with timed input

10.1: Introduction. Cell types in neural tissue: Neurons Neuroglial cells (also known as neuroglia, glia, and glial cells) Dendrites.

Physiology Unit 2 CONSCIOUSNESS, THE BRAIN AND BEHAVIOR

Direct memory access using two cues: Finding the intersection of sets in a connectionist model

How Synapses Integrate Information and Change

Active Control of Spike-Timing Dependent Synaptic Plasticity in an Electrosensory System

Transcription:

19th AWCBR (Australian Winter Conference on Brain Research), 21, Queenstown, AU https://www.otago.ac.nz/awcbr/proceedings/otago394614.pdf Do local modification rules allow efficient learning about distributed representations? A. R. Gardner-Medwin Dept. Physiology, University College London, London WC1E6BT, UK Distributed representations permit very many distinguishable events to be coded on a set of cells, with each cell used in many events. Since synaptic modifications can only depend on local influences, there is a fundamental problem learning how often and under what conditions distributed patterns of activity may occur. Frequencies of use and the associations of individual active elements can be measured locally and pooled for active elements within an event, but overlap leads to interference that can only be compensated on an average basis, with inevitable added variance. This constrains the compactness of distributed representations if they are to operate efficiently. Gardner- Medwin & Barlow (21, Neural Computation 13: 477-54) have employed counting (a form of familiarity discrimination) to explore such constraints quantitatively. underlies estimation of probabilities and association, and is fundamental to learning. Though precise counts are only possible if events have direct representations, high (i.e. effective use of available data samples) is only possible with distributed representations at the cost of high redundancy. With counts based on usage of individual cells, efficiencies >5% require a number of cells (Z) at least comparable to the number of events (N). Synapses counting activity in pairs of cells can reduce this to around Z=(6N).5 cells. More sophisticated use of dendritic information (e.g. modification conditions requiring combined activity in 3 cells: adjacent presynaptic terminals and a post-synaptic cell) can improve performance, but nowhere near the information theoretic limit: Z=log 2 (N).

DO LOCAL MODIFICATION RULES ALLOW EFFICIENT LEARNING ABOUT DISTRIBUTED REPRESENTATIONS? A. R. Gardner-Medwin THE PRINCIPLE OF LOCAL COMPUTABILITY Neural computations must satisfy a constraint - each modifiable parameter is a function only of the history of locally available information. Realistic models must be capable of expression in this form - without functions of global patterns. In synaptic terms : modification can only depend on the history of pre- and postsynaptic membrane and cytoplasm states and local extracellular modulators. DISTRIBUTED REPRESENTATIONS Events (significant for learning & inference) correspond to spatial patterns - sets of active cells - and are not usually the sole trigger features for individual cells. Distributed representations can be compact (representing many distinct events on few cells), but... THE OVERLAP PROBLEM Each cell and synapse is involved in experience of many events, causing interference and inevitable errors in learning and statistics for individual events. THESE CONSIDERATIONS SEEM TO ENFORCE THIS BROAD STRATEGY :- Measure statistics for individual cells and synapses Combine for all cells or synapses active in an event Correct for the average interference due to overlap Tolerate the variance introduced Questions: How can we quantify the variance that is tolerable? - the concept of. What are the constraints on the number of events that can be learned about, on a given number of cells? What are the implications for the nature of distributed representations and trigger features, for efficient learning?

d Number Events - a challenge for distributed representations With H. Barlow (21, Neural Computation 13: 477-54) we examined the problem of counting events. We argue that unless a representation can sustain efficient counting of events, it cannot sustain efficient learning. Why Count? But... neural systems do not need to count precisely Biological success is about prediction Prediction is about conditional probabilities Probabilities come from frequencies Frequencies require counting But NB typically: Actual N = expected value m m Note that what matters is the estimate of m, not the value of N This implies that there is little point in counting much more accurately than m (roughly, N) Statistical of an estimation (Fisher) Efficiency = data needed for a given reliability, with a perfect (counting) algorithm data needed when there is added variance 5% it takes twice as long to achieve reliable inferences a Poisson process with variance = V :- Efficiency = % x m / (m + V) 6 Estimating the count of randomly scattered X s When forced to subitise (i.e. to estimate a count without sequential matching to linguistic symbols, fingers, etc.) human counting falls with large numbers, but is typically >5% for counts in the spatial domain. 5 4 3 2 N ± N 2 3 4 5 6 Displayed Number

Efficiency for events Dept. of Physiology, University College London, London WC1E 6BT, UK Models & simulation examples of performance based on usage of (A) cells, (B) pre- & post-synaptic cell pairings and (C) triplets of cells A) Projection model : sums weights proportional to usage of cells active in a test event. B) Support model : measure auto-associative support for activity in a test event using imposed inhibition. Synapse weights amount of pre- and postsynaptic paired activity. C) Triplet support As B, but with synapses that strengthen with cooccurrences of activity in 3 cells: pre- and post- plus an adjacent pre- axon. Under test conditions (recall) active synapses (red) sum their strengths, but only when adjacent pairs are active (a), not single synapses (b,c). a. 2 - A) =% 5-5 b. 5 B) =65% 5-5 c. 5 C) =91% Simulation results (shown above) with the 3 models:- 2 distinct events were represented by random selections of binary active cells on a network of cells Events occurred differing numbers of times shown on the horizontal axes (Poisson distributed, mean=4) Simulated estimates of the number of occurrences are plotted on the vertical axes, from activation on test presentation of each event. For (C) cells had 2 afferent synapses from each other cell, shuffled so that neighbours were never identical, but otherwise random. Distributed representations can support high, but they must employ very many cells compared with a compact representation for the counted events..5 1 Support Model (B) a =.1 a = optimal Projection Model (A) Total number of cells (Z) NB cells would allow each event to have a direct representation, giving % counting. 7 cells suffice for a compact representation distinguishing events. a (the activity ratio) is the fraction of cells active in each event.

How can better representations improve? 1. Use more cells & trigger features than there are events of interest E.g. counting equiprobable events:- Efficiency Efficiency active total (projection (support number of cells cells model - A) model -B) distinct reps 12 3 22% 45% 8 4 3 22% 63% 4 5 5% 93% 8 4 3 75% 98% 8 Direct Reps: 1 % - 2 1 8 % - 8 NB distributed reps handle almost any of the many distinct events. Direct reps must be set up in advance ( restricted to familiar events) 2. Vary the number of active cells between events Rare events tend to have poor counting because of overwhelming interference from overlap with common events. Increase the number of cells active in rare (and important) events, e.g. through alerting & orienting reactions and selective attention. Reduce the number of cells active in common events through habituation, adaptation. No. of Active Cells 1.1 1 Expected no. of occurrences of individual events % Efficiency 5%.1 1 Expected no. of occurrences of individual events 3. Choose trigger features appropriately Trigger features (i.e. the conditions that lead to activity of individual cells) should ideally be such that :- there is minimum overlap between representations of different events, especially ones that have different significance in relation to learning. the subset of events in which a feature is active tend to have similar associations, and significance in relation to learning. Conclusions - Distributed vs. Direct representations Distributed representations can be compact (requiring few cells), but do not permit % counting. Distributed reps do permit high efficiencies (>>5%) but only if they are highly redundant (comparable to or greater than direct reps). Distributed reps allow novel events to be counted without prior provision for special encoding, even when the interest in a type of event emerges only after the experience. In this context distributed reps have a huge advantage over direct reps. Distributed reps permit the boosting of counting efficiencies for rare or significant events by variation of activity ratios through mechanisms such as attention and habituation. Direct reps allow the processing of simultaneous events in parallel. Distributed reps require that simultaneous events be processed serially, as in a stream of conscious awareness. Serial processing of simultaneous events