Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention

Similar documents
Computational Explorations in Cognitive Neuroscience Chapter 7: Large-Scale Brain Area Functional Organization

NEOCORTICAL CIRCUITS. specifications

Introduction to Computational Neuroscience

Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence

Sparse Coding in Sparse Winner Networks

ERA: Architectures for Inference

Computational Cognitive Neuroscience

Introduction to Computational Neuroscience

Cognitive Neuroscience Section 4

Introduction to Computational Neuroscience

Human Cognitive Developmental Neuroscience. Jan 27

PHY3111 Mid-Semester Test Study. Lecture 2: The hierarchical organisation of vision


Overview of the visual cortex. Ventral pathway. Overview of the visual cortex

Basics of Perception and Sensory Processing

Memory, Attention, and Decision-Making

Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6)

Neuroinformatics. Ilmari Kurki, Urs Köster, Jukka Perkiö, (Shohei Shimizu) Interdisciplinary and interdepartmental

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

Identify these objects

PSY 310: Sensory and Perceptual Processes 1

Touch PSY 310 Greg Francis. Lecture 34. Touch receptors

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J.

On the implementation of Visual Attention Architectures

Rolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.

The Motor Systems. What s the motor system? Plan

Self-Organization and Segmentation with Laterally Connected Spiking Neurons

Objectives. Objectives Continued 8/13/2014. Movement Education and Motor Learning Where Ortho and Neuro Rehab Collide

Mechanosensation. Central Representation of Touch. Wilder Penfield. Somatotopic Organization

A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences

using deep learning models to understand visual cortex

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Lecture 2.1 What is Perception?

PSY380: VISION SCIENCE

Computational & Systems Neuroscience Symposium

Supplementary Motor Area exerts Proactive and Reactive Control of Arm Movements

Just One View: Invariances in Inferotemporal Cell Tuning

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

Toward A Cognitive Computer Vision System

Artificial Intelligence

Realization of Visual Representation Task on a Humanoid Robot

DYNAMICISM & ROBOTICS

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

Uniformity of the neocortex is the basis for human versa4lity

KINE 4500 Neural Control of Movement. Lecture #1:Introduction to the Neural Control of Movement. Neural control of movement

Lecture 6. Perceptual and Motor Schemas

Timing and the cerebellum (and the VOR) Neurophysiology of systems 2010

Neurophysiology and Information

Learning Classifier Systems (LCS/XCSF)

(Visual) Attention. October 3, PSY Visual Attention 1

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

Memory Prediction Framework for Pattern Recognition: Performance and Suitability of the Bayesian Model of Visual Cortex

Neural Cognitive Modelling: A Biologically Constrained Spiking Neuron Model of the Tower of Hanoi Task

Brain Mechanisms Explain Emotion and Consciousness. Paul Thagard University of Waterloo

Zoo400 Exam 1: Mar 25, 1999

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

Consciousness as representation formation from a neural Darwinian perspective *

A Multimodal Paradigm for Investigating the Perisaccadic Temporal Inversion Effect in Vision

A Three Agent Model of Consciousness Explains Spirituality and Multiple Nondual Enlightened States Frank Heile, Ph.D.

5/20/2014. Leaving Andy Clark's safe shores: Scaling Predictive Processing to higher cognition. ANC sysmposium 19 May 2014

Neocortex. Cortical Structures in the Brain. Neocortex Facts. Laminar Organization. Bark-like (cortical) structures: Shepherd (2004) Chapter 12

R Jagdeesh Kanan* et al. International Journal of Pharmacy & Technology


Hierarchical dynamical models of motor function

Probabilistic Models of the Cortex: Stat 271. Alan L. Yuille. UCLA.

Consciousness The final frontier!

Artificial Intelligence Lecture 7

Lateral Inhibition Explains Savings in Conditioning and Extinction

Object recognition and hierarchical computation

CISC 3250 Systems Neuroscience

Neural Cognitive Modelling: A Biologically Constrained Spiking Neuron Model of the Tower of Hanoi Task

Learning in neural networks

Bioscience in the 21st century

Degree of freedom problem

M.Sc. in Cognitive Systems. Model Curriculum

Spatial Attention: Unilateral Neglect

Rajeev Raizada: Statement of research interests

ISSN (PRINT): , (ONLINE): , VOLUME-5, ISSUE-5,

Object vision (Chapter 4)

Implementation of Perception Classification based on BDI Model using Bayesian Classifier

Cognitive Penetrability and the Content of Perception

Sensory Cue Integration

A neural circuit model of decision making!

Competing Frameworks in Perception

Competing Frameworks in Perception

Grounded Cognition. Lawrence W. Barsalou

Sleep-Wake Cycle I Brain Rhythms. Reading: BCP Chapter 19

The issue (1) Some spikes. The issue (2) Am I spikes over time?

Oxford Foundation for Theoretical Neuroscience and Artificial Intelligence

Modelling Decision Making under Uncertainty Machine Learning and Neural Population Techniques. Elaine Duffin

Time Experiencing by Robotic Agents

Principals of Object Perception

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

EDGE DETECTION. Edge Detectors. ICS 280: Visual Perception

The Science. This work was performed by Dr. Psyche Loui at Wesleyan University under NSF-STTR # Please direct inquiries to

A Three Agent Model of Consciousness Explains Spirituality and Multiple Nondual Enlightened States Frank Heile, Ph.D.

An Escalation Model of Consciousness

Evaluating the Effect of Spiking Network Parameters on Polychronization

Cerebral Cortex. Edmund T. Rolls. Principles of Operation. Presubiculum. Subiculum F S D. Neocortex. PHG & Perirhinal. CA1 Fornix CA3 S D

Transcription:

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention Tapani Raiko and Harri Valpola School of Science and Technology Aalto University (formerly Helsinki University of Technology)

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention Tapani Raiko and Harri Valpola ZenRobotics Ltd. www.zenrobotics.com

Computational neuroscience group What: Figuring out how the brain works. How: Building brains for robots = system-level modelling, implementing a whole vertebrate/mammalian brain. Why: Because we can. Intact cerebellum and sober Cerebellar lesion or drunk

Our main strengths in research Neuroscience machine learning / AI / neural nets robotics in unstructured environments Cognitive architecture: the organisation of the whole brain Cerebral cortex Basal ganglia Cerebellum Hippocampus

ZenRobotics Ltd. Sorting waste with intelligent robots

ZenRobotics Ltd. Waste recycling lab

ZenRobotics Ltd. Robot s viewpoint

Cerebral cortex Modelling the world (and yourself as part of it) Forward modelling: prediction and simulation Inverse modelling: figuring out which actions lead to desired consequences; planning

Different types of inputs and outputs Primary input usually from bottom-up (from the senses) Numerous feedback connections (order of 10 x)

Introduction: Bayesian brain Generative model / Bayesian brain hypothesis The generative model allows to infer the causes of sensations and to predict inputs Hierarchical Cortical structure Top-down priors (context) Perception = Probabilistic integration of information using Bayesian inference Bottom-up sensory data From the presentation of Salvador Dura, BICS 2010, 14.7.2010

Bayesian theory says: Decisions are based on 1. Beliefs (measured by probability) 2. Utilities The recipe: 1. Evaluate the probabilities of all possible states of the world (probabilistic inference) 2. Evaluate the probabilities of all outcomes for each and every potential action (probabilistic inference) 3. Choose the action which maximises the expected utility This is optimal if there are no restrictions on the available computational resources

Selection of information But computational resources are restricted It is impossible to consider all the states and actions It is necessary to select information in order to make decisions

Selection of information In practice, it has turned out to be impossible to learn complex abstractions from real data bottom-up There is too much structure it is necessary to select which abstractions (groupings of elementary features) are meaningful Information will be lost Example: learning the phonemes results in inability to distinguish between foreign phonemes

Cerebral cortex Modelling the world (and yourself as part of it) Forward modelling: prediction and simulation Inverse modelling: figuring which actions lead to desired consequences; planning Selection of useful information Selective learning of features: selecting useful high-level abstractions (sensory and motor) Selective attention Modelling objects and their relations Segmentation of objects

Key problem: How to select useful information? It is necessary to select information in order to make decisions Selection is a type of decision, in other words: In order to decide we need to decide Infinite regress!

Distributed selection on cortex Primary input usually from bottom-up (from the senses) Context (top-down or lateral) guides selection both in learning and on behavioural timescale

Invariant features Group simple features into complex ones in a hierarchical model What is the criterion? Slow feature analysis: features that are activated close-by in time Subspace ICA: features that are activated together Canonical correlation analysis: features that are activated in the same context

Canonical correlation analysis A statistical technique which finds what is in common between two sets of data Find two projections (one from each dataset) such that their correlation is maximized Generalizes to several datasets, nonlinearities E.g., find visual features which are most relevant for motor control On behavioural timescale, activations are determined mainly by visual bottom-up inputs Motor context guides learning

Cortical long-range connections are specific Inhibitory connections (white dots) are local and symmetric Long-range excitatory connections (black dots) adapt through experience Kevan Martin, Current Opinion in Neurobiology, 2002

Context-guided learning of features Inputs are whitened Context (top-down and lateral) biases bottom-up activations Competitive learning invariant features emerge (e.g., complex cells)

Attentional modulation of competition A V4 neuron is recorded Weak activity for house, strong activity for face Intermediate activity for a combination Excitation adds up but so does inhibition Selective attention can mask the effect of the distractor Reynolds and Chelazzi, Annual Review of Neuroscience, 2004

Biased competition model Local competition on each cortical area Context (top-down and lateral) biases the competition Selective attention emerges from the dynamics primary input contextual input local inhibition

Visual search Top-down biasing from working memory implements visual search May look like sequential search (time increases with distractors) but the mechanism is fully parallel Both bottom-up and topdown phenomena related to (covert) attention can be explained Deco and Rolls, Vision Research, 2004

Attention and learning: selection on different timescales Within the Bayesian framework, the only difference between perceptual inference and learning is the timescale In adults, perceptual learning is very strongly dependent on attention E.g., the same bottom-up input but different tasks learn to perceive different aspects Selection in both attention and learning, only timescales differ

Biased competitive learning Biased competition + competitive learning (Master s thesis of Antti Yli-Krekola, 2007) Add adaptation (habituation, getting tired) attention switches between objects Learning can be dramatically improved by switching attention (our paper at ICANN 2009)

Problems with engineering solutions to segmentation Many engineering solutions suffer from a chicken-or-egg problem: Recognition is usually successful only after segmentation Segmentation is often successful only after recognition Iterative bottom-up / top-down message passing solves this problem in biased competition model

Segmentation still often remains poor Although biased competition will select one object at a time, segmentation can be poor From the viewpoint of an individual neuron/feature, we are effectively asking: Do you belong to the currently active object or not? An easier question would be: Do you belong to object1 or object2 or or something else? In Bayesian terms: explaining away

Segmentation and synchrony Engel, Fries and Singer, Nature Reviews Neuroscience, 2001 Hypothesis: neurons encoding the same object synchronize

Segmentation with weakly coupled oscillators: LEGION model http://www.scholarpedia.org/article/legion:_locally_excitatory_globally_inhibitory_oscillator_networks

Biased competitive learning + coupled oscillators Our goal in BICS 2010 paper is to study how biased competitive learning could be combined with coupled oscillators for better segmentation Abstraction level: The behaviour of one area is what we design try to keep it as simple as possible Try to come up with emergent oscillatory segmentation in a network of interconnected areas

Desired emergent properties 1. Integrates information from local patches of each object translates into the gain of the population 2. The objects with globally highest gain emerge 3. Objects synchronise completely (not pieces of objects) 4. Objects desynchronise between each other (when their representations overlap)

Properties of a single area part 1 Pretty much like biased competitive learning but: Intrinsic oscillators built from excitatory and inhibitory neurons Low-pass filtering is needed somewhere to build oscillators. Low-pass filter inhibition to avoid distorting the signal carried by excitatory neurons. Keep activations in check with gain control

Properties of a single area part 2 If no bottom-up input no activation If bottom-up input If no support little oscillation, little activation If constant support some oscillation, more activation If oscillating support phase-locked oscillation, strong activation Competing features push each others phases further away from each other Details are not crucial, there are many ways to implement these properties

Results part 1 First we checked that our implementation got the properties for a single area right (using externally generated bottom-up and lateral input) Seemed to work ok What happens now when we connect many areas together?

Results part 2 Use simple visual inputs A network of 5 5 inputs Learn bottom-up features and their lateral connections (between excitatory neurons)

Results part 3 50 55 60 65 70 75 80 85 90 95 100 105 110 115 120 125 130 135 140 145 150 155 160 165 170 175 180 185 190 195 200 205 210 215 220 225 230 235 240 245 250 255 260 265 270 275 280 285 290

Looks promising Discussion There was a problem with multiple activations from the same object within one area Hierarchy and top-down support would help tie them together? Complex cells? Synchronize inputs? Obvious next step: test a hierarchical model But I want to see it work in motor control: Is this really useful? Even context-guided learning and biased competition haven t been properly integrated with motor learning test them separately first

Future / ongoing work Sensory and motor abstractions Development of sensory abstractions guided by motor context and vice versa (akin to canonical correlation analysis) Predictive power as a measure of value of information (needed for selection of information) Decision-making as biased competition on motor representations Better mappings A hierarchical model of correlations at lower levels; latent variables describing the operating points of the system

Conclusions The experiments with artificial data have not proven that the system works in real life Nevertheless, looks promising I expect to revisit this work: once we need sophisticated segmentation, synchrony really might be a useful ingredient

Thank you! www.zenrobotics.com 7/15/2010