Sparse Coding in Sparse Winner Networks

Similar documents
Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence

Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6)

Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells

Spiking Inputs to a Winner-take-all Network

Rolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.

Intelligent Control Systems

Fundamentals of Computational Neuroscience 2e

A Neurally-Inspired Model for Detecting and Localizing Simple Motion Patterns in Image Sequences

Realization of Visual Representation Task on a Humanoid Robot

ERA: Architectures for Inference

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention

Cell Responses in V4 Sparse Distributed Representation

CRISP: Challenging the Standard Framework of Hippocampal Memory Function

Introduction to Computational Neuroscience

Learning in neural networks

lateral organization: maps

Auto-Encoder Pre-Training of Segmented-Memory Recurrent Neural Networks

Computational Cognitive Neuroscience

Lateral Inhibition Explains Savings in Conditioning and Extinction

International Journal of Scientific & Engineering Research Volume 4, Issue 2, February ISSN THINKING CIRCUIT

On Intelligence. Contents. Outline. On Intelligence

University of Cambridge Engineering Part IB Information Engineering Elective

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos

Lecture 2.1 What is Perception?

Why do we have a hippocampus? Short-term memory and consolidation

Applied Neuroscience. Conclusion of Science Honors Program Spring 2017

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

Thalamocortical Dysrhythmia. Thalamocortical Fibers. Thalamocortical Loops and Information Processing

CHAPTER I From Biological to Artificial Neuron Model

Modeling of Hippocampal Behavior

Challenges of Embodied Intelligence

DeSTIN: A Scalable Deep Learning Architecture with Application to High-Dimensional Robust Pattern Recognition

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

Noise Cancellation using Adaptive Filters Algorithms

NEOCORTICAL CIRCUITS. specifications

Recognition of English Characters Using Spiking Neural Networks

Katsunari Shibata and Tomohiko Kawano

Why is dispersion of memory important*

Type II Fuzzy Possibilistic C-Mean Clustering

Course Introduction. Neural Information Processing: Introduction. Notes. Administration

Intelligent Machines That Act Rationally. Hang Li Bytedance AI Lab

Intelligent Machines That Act Rationally. Hang Li Toutiao AI Lab

Learning and Adaptive Behavior, Part II

TIME SERIES MODELING USING ARTIFICIAL NEURAL NETWORKS 1 P.Ram Kumar, 2 M.V.Ramana Murthy, 3 D.Eashwar, 4 M.Venkatdas

Active Sites model for the B-Matrix Approach

LECTURE 5: REACTIVE AND HYBRID ARCHITECTURES

COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks

NMF-Density: NMF-Based Breast Density Classifier

Neural Information Processing: Introduction

AU B. Sc.(Hon's) (Fifth Semester) Esamination, Introduction to Artificial Neural Networks-IV. Paper : -PCSC-504

Predicting Breast Cancer Survivability Rates

Hebbian Plasticity for Improving Perceptual Decisions

using deep learning models to understand visual cortex

Flexible, High Performance Convolutional Neural Networks for Image Classification

Computational Explorations in Cognitive Neuroscience Chapter 3

Physiology of Tactile Sensation

Understanding Convolutional Neural

A general error-based spike-timing dependent learning rule for the Neural Engineering Framework

Memory, Attention, and Decision-Making

Artificial Neural Networks to Determine Source of Acoustic Emission and Damage Detection

Observational Learning Based on Models of Overlapping Pathways

PSY380: VISION SCIENCE

Neuromorphic computing

Lecturer: Dr. Benjamin Amponsah, Dept. of Psychology, UG, Legon Contact Information:

A model of the interaction between mood and memory

Genetically Generated Neural Networks I: Representational Effects

Biceps Activity EMG Pattern Recognition Using Neural Networks

Self-Organization and Segmentation with Laterally Connected Spiking Neurons

Cerebral Cortex. Edmund T. Rolls. Principles of Operation. Presubiculum. Subiculum F S D. Neocortex. PHG & Perirhinal. CA1 Fornix CA3 S D

Networks and Hierarchical Processing: Object Recognition in Human and Computer Vision

STDP-based spiking deep convolutional neural networks for object recognition

Emergent creativity in declarative memories

A hybrid Model to Estimate Cirrhosis Using Laboratory Testsand Multilayer Perceptron (MLP) Neural Networks

Where-What Network 3: Developmental Top-Down Attention for Multiple Foregrounds and Complex Backgrounds

16/05/2008 #1

Using stigmergy to incorporate the time into artificial neural networks

Predicting Breast Cancer Recurrence Using Machine Learning Techniques

Convolutional Spike Timing Dependent Plasticity based Feature Learning in Spiking Neural Networks

Neuromorphic convolutional recurrent neural network for road safety or safety near the road

A Hubel Wiesel Model of Early Concept Generalization Based on Local Correlation of Input Features

Neural Coding. Computing and the Brain. How Is Information Coded in Networks of Spiking Neurons?

NERVOUS SYSTEM 1 CHAPTER 10 BIO 211: ANATOMY & PHYSIOLOGY I

SMARTfit Multisensory Learning Develops the Brain and the Body. A 30 minute PE class on SMARTfit helps the brain to learn better

Application of Tree Structures of Fuzzy Classifier to Diabetes Disease Diagnosis

Prior Knowledge and Memory Consolidation Expanding Competitive Trace Theory

Improving Associative Memory in a Network of Spiking

ADAPTING COPYCAT TO CONTEXT-DEPENDENT VISUAL OBJECT RECOGNITION

Lesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS

Welcome to CSE/NEUBEH 528: Computational Neuroscience

Errol Davis Director of Research and Development Sound Linked Data Inc. Erik Arisholm Lead Engineer Sound Linked Data Inc.

Artificial Intelligence Lecture 7

Thalamocortical Feedback and Coupled Oscillators

Representational Switching by Dynamical Reorganization of Attractor Structure in a Network Model of the Prefrontal Cortex

Image Enhancement and Compression using Edge Detection Technique

Classification of benign and malignant masses in breast mammograms

NEURAL SYSTEMS FOR INTEGRATING ROBOT BEHAVIOURS

Dendritic compartmentalization could underlie competition and attentional biasing of simultaneous visual stimuli

196 IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, VOL. 2, NO. 3, SEPTEMBER 2010

Position invariant recognition in the visual system with cluttered environments

Transcription:

Sparse Coding in Sparse Winner Networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University, Athens, OH 45701 {starzyk, yliu}@bobcat.ent.ohiou.edu 1 2 Ross University School of Medicine Commonwealth of Dominica dvogel@rossmed.edu.dm Abstract. This paper investigates a mechanism for reliable generation of sparse code in a sparsely connected, hierarchical, learning memory. Activity reduction is accomplished with local competitions that suppress activities of unselected neurons so that costly global competition is avoided. The learning ability and the memory characteristics of the proposed winner-take-all network and an oligarchy-take-all network are demonstrated using experimental results. The proposed models have the features of a learning memory essential to the development of machine intelligence. Keywords. sparse coding, winner-takes-all, Hebbian learning in sparse structure, hierarchical self-organizing memory 1 Introduction In this paper we describe a learning memory built as a hierarchical, self-organizing network in which many neurons activated at lower s represent detailed features, while very few neurons activated at higher s represent objects and concepts in the sensory pathway [1]. By recognizing the distinctive features of patterns in a sensory pathway, such a memory may be made to be efficient, fault-tolerant, and to a useful degree, invariant. Lower features may be related to multiple objects represented at higher s. Accordingly, the number of neurons increases up the hierarchy with the neurons at lower s making divergent connections with those on higher s [2]. This calls to mind the expansion in number of neurons along the human visual pathway (e.g., a million geniculate body neurons drive 200 million V1 neurons [3]). Self-organization is a critical aspect of the human brain in which learning occurs in an unsupervised way. Presentation of a pattern activates specific neurons in the sensory pathway. Gradually, neuronal activities are reduced at higher s of the hierarchy, and sparse data representations, usually referred to as sparse codes, are built. The idea of sparse coding emerged in several earlier works [4][5]. In recent years, various experimental and theoretical studies have supported the assumption that information in real brains is represented by a relatively small number of active neurons out of a large neuronal population [6][[7][3]. In this paper, we implement the novel idea of performing pathway selections in sparse network structures. Self-organization and sparse coding are obtained by means

of localized, winner-take-all (WTA) competitions and Hebbian learning. In addition, an oligarchy-take-all (OTA) concept and its mechanism is proposed that produces redundant, fault tolerant, information coding. This paper is organized as follows. In section 2, a winner network is described that produces sparse coding and activity reduction in the learning memory. In section 3, an OTA network is described that produces unsupervised, self-organizing, learning with distributed information representations. Section 4 demonstrates the learning capabilities of the winner and the OTA networks using experimental results. Finally, our method of sparse coding in sparse structures is summarized in section 5. 2 The Winner Network In the process of extracting information from data, we expect to predictably reduce neuronal activities at each of a sensory pathway. Accordingly, a competition is required at each. In unsupervised learning, we need to find the neuron in the network that has the best match to the input data. In neural networks, such a neuron, is usually determined using a WTA network [8][9]. A WTA network is usually implemented based on competitive neural network in which inhibitory lateral links and recurrent links are utilized, as shown in Fig. 1. The outputs iteratively suppress the signal strength among each other and the neuron with maximum signal strength will stay as the only active neuron when the competition is done. For a large memory, with many neurons on the top, a global WTA operation is complex, inaccurate and costly. Moreover, average competition time increases as the likelihood of similar signal strengths increases in large WTA networks. d1 dj dm -α -α 1 1 1 -α r1 rj rm Fig. 1. WTA network as competitive neural network The use of sparse connections between neurons can, at the same time, improve efficiency and reduce energy consumption. However, sparse connections between neurons on different hierarchical s may fail to transmit enough information along the hierarchy for reliable feature extraction and pattern recognition. In a local network model for cognition, called an R-net [10][11], secondary neurons, with random connections to a fraction of primary neurons in other layers, effectively provide almost complete connectivity between primary neurons pairs. While R-nets provide large capacity, associative memories, they were not used for feature extraction and sparse coding in the original work. The R-net concept is expanded in this work by using secondary neurons to fully connect primary neurons on lower s to primary neurons on higher s through

the secondary neurons of a sparsely connected network. The network has an increasing number of neurons on the higher s, and all neurons on the same have an equal number of input links from neurons on the lower. The number of secondary s between primary s affects the overall network sparsity. More secondary s can be used to increase the network sparsity. Such sparsely connected network with secondary s is defined as winner network and illustrated in Fig. 2. Increasing number of Overall neurons Primary h+1 Secondary s Primary h Fig. 2. Primary and secondary in winner network The initial random input weights to each neuron are scaled to have a sum of squared weights equal to 1, which places them on the unit multidimensional sphere. Because a neuron becomes active when its input weight vector is similar to its input pattern, spreading the input weights uniformly on the unit-sphere increases the memory capacity of the winner network. Furthermore, the normalization of the weights maintains the overall input signal so that the output signal strength of neurons, and accordingly the output of the network, will not be greatly affected by the number of input connections. In a feed-forward computation, each neuron combines its weighted inputs using a thresholded activation function. Only when the signal strength is higher than the activation threshold can the neuron send a signal to its post-synaptic neurons. Eventually, the neurons on the highest will have different s of activation, and the most strongly activated neuron (the global winner) is used to represent the input pattern. In this work, the competition to find the global winner is replaced by small-scale WTA circuits in local regions in the winner network as described next. In a sparsely connected network, each neuron on the lower connects to a group of neurons on the next higher. The winning neuron at this is found by comparing neuronal activities. In Hebbian learning, weight adjustments reduce the plasticity of the winning neuron s connections. Therefore, a local winner should not only have the maximum response to the input, but also its connections should be flexible enough to be adjusted towards the input pattern so that the local winner is, + 1 swinner i = max w ( 1,2,.. ) 1 jk sk ji i = N ρ, (1) + j N i k N j where N +1 i is a set of post-synaptic neurons on (+1) driven by a neuron i, N i is a set of pre-synaptic neurons that project onto neuron j on (), and ρ ji 3

denotes the plasticity of the link between pre-synaptic neuron i and post-synaptic neuron j, as shown in Fig. 3(a). j +1 N i 1 2 3 4 5 6 7 8 9 1 2 3 4 5 i N j 6 7 +1 S winner S winner S winner S winner S winner Winner network Loser neurons in local competition Winner neurons in local competition Fig. 3. (a) Interconnection structure to determine a local winner, (b) The winner network Such local competition can be easily implemented using a current-mode WTA circuit [12]. A local winner neuron, for example N +1 4 in Fig. 3(a), will pass its signal strength to its pre-synaptic neuron N 4, and all other post-synaptic branches connecting neuron N 4 with the losing nodes will be logically cut off. Such local competition is done first on the highest. The signal strengths of neurons which win in their corresponding local competitions propagate down to the lower s and the same procedure continues until the first input layer is reached. The global winning neuron on the top depends on the results of all local competitions. Subsequently, the signal strength of the global winner is propagated down to all lower- neurons which connect to the global winner. Most of the branches not connected to the global winner are logically cut off, while the branches of the global winner are kept active. All the branches that propagate the local winner signal down the hierarchy form the winner network, as shown in Fig. 3(b). Depending on the connectivity structure, one or more winner networks can be found. By properly choosing the connectivity structure, we may guarantee that all of the input neurons are in a single winner network so that the output contains a single winner. Let us use a 3-layer winner network (1 input, 2 secondary s and 1 output ) as an example. The network has 64 primary input neurons and 4096 output neurons with 256 and 1024 secondary neurons, respectively. The number of active neurons in the top decreases with increasing numbers of input connections. As shown in Fig.4, when the number of input links to each neuron is more than 8, a single winner neuron in the top is achieved. Since the branches logically cut off during local competition will not contribute to post-synaptic neuronal activities, the synaptic strengths are recalculated only for branches in the winner network. As all the branches of the winner network are used, the signal strength of pathways to the global winner are not reduced. However, due to the logically disconnected branches, the signal strength of pathways to other output neurons are suppressed. As a result, an input pattern activates only some of the neurons in the winner networks. The weights are only adjusted using Hebbian learning for links in winner networks to reinforce the activation of the global winner. After updating, weights are scaled so that they are still spread on the unit-sphere.

In general, the winner network with secondary neurons and sparse connections, builds sparse representations in three steps: sending data up through the hierarchy, finding the winner network and global winner by using local competitions, and training. The winner network finds the global winner efficiently without iterations usually adopted in MAXNET [8][9]. It provides an effective and efficient solution to the problem of finding global winners in large networks. The advantages of sparse winner networks are significant for large size memories. Number of active neurons on top vs. Number of input links to each neuron 12 number of active neurons on top 10 8 6 4 2 0 2 3 4 5 6 7 8 9 10 number of input links Fig. 4. Effect of number of input connections to neurons 3 Winner Network with Oligarchy-Takes-All The recognition using a single-neuron representation scheme in the winner network can easily fail because of noise, fault, variant views of the same object, or learning of other input patterns due to an overlap between activation pathways. In order to have distributed, redundant data representations, an OTA network is proposed in this work to use a small group of neurons as input representations. In an OTA network, the winning neurons in the oligarchy are found directly in a feed-forward process instead of the 3-step procedure used in the winner network as described in section 2. Neurons in the 2 nd layer combine weighted inputs and use a threshold activation function as in the winner network. Each neuron in the 2 nd layer competes in a local competition. The projections onto losing nodes are logically cut off. The same Hebbian learning as is used in the winner network is carried out on the logically connected links. Afterwards, the signal strengths of the 2 nd are recalculated considering only effects of the active links. The procedure is continued until the top of hierarchy is reached. Only active neurons on each are able to send the information up the hierarchy. The group of active neurons on the top provides redundant distributed coding of the input pattern. When similar patterns are presented, it is expected that similar groups of neurons will be activated. Similar input patterns can be recognized from the similarities of their highest representations. 5

4 Experimental results The learning abilities of the proposed models were tested on the 3-layer network described in section 2. The weights of connections were randomly initialized within the range [-1, 1]. A set of handwritten digits from the benchmark database [13] containing data in the range [-1, 1] was used to train the winner network or OTA networks. All patterns have 8 by 8 grey pixel inputs, as shown in Fig. 5. Each input pattern activates between 26 and 34 out of 4096 neurons on the top. The groups of active neurons in the OTA network for each digit are shown in Table 1. On average, each pattern activates 28.3 out of 4096 neurons on the top with the minimum number of 26 neurons and the maximum number of 34 neurons. Fig. 5. Ten typical patterns for each digit Table 1. Active neuron index in the OTA network for handwritten digit patterns digit Active Neuron index in OTA network 0 72 91 365 371 1103 1198 1432 1639 1 237 291 377 730 887 1085 1193 1218 2 294 329 339 771 845 1163 1325 1382 3 109 122 237 350 353 564 690 758 4 188 199 219 276 307 535 800 1068 5 103 175 390 450 535 602 695 1008 6 68 282 350 369 423 523 538 798 7 237 761 784 1060 1193 1218 1402 1479 8 35 71 695 801 876 1028 1198 1206 9 184 235 237 271 277 329 759 812 The ability of the network to classify was tested by changing 5 randomly selected bits of each training pattern. Comparing the OTA neurons obtained during training with those activated by the variant patterns, we find that the OTA network successfully recognizes 100% of the variant patterns. It is expected that changing more bits of the original patterns will degrade recognition performance. However, the tolerance of the OTA network for such change is expected to be better than that of the winner network. Fig. 6 compares the performances of the winner network and the OTA network for different numbers of changed bits in the training patterns based on 10 Monte-Carlo trials. We note that increasing the number of changed bits in the patterns quickly degrades the winner network s performance on this recognition task. When the number of bits changed is larger than 20, the recognition correctness stays around 10%. However,10% is the accuracy for random recognition for 10 digit patterns recognition. It means that when the number of changed bits is over 20, the winner network is not able to make useful recognition. As anticipated, the OTA network has much better fault tolerance and it is resistant to this degradation of recognition correctness.

percentage of correct recognition 1 0.8 0.6 0.4 Percentage of correct recognition performance of OTA performance of winner network 0.2 Accuracy of random recognition 0 0 10 20 30 40 50 number of bits changed in the pattern Fig. 6. Recognition performance of the OTA network and the winner network 5. Conclusions This paper investigates a mechanism for reliably producing sparse coding in sparsely connected networks and building high capacity memory with redundant coding into sensory pathways. Activity reduction is accomplished with local rather than global competition, which reduces hardware requirements and computational cost of self-organizing learning. High memory capacity is obtained by means of layers of secondary neurons with optimized numbers of interconnections. In the winner network, each pattern activates a dominant neuron as its representation. In the OTA network, a pattern triggers a distributed group of neurons. With OTA, information is redundantly coded so that recognition is more reliable and robust. The learning ability of the winner network is demonstrated using experimental results. The proposed models produce features of a learning memory that may prove essential for developing machine intelligence. Reference: 1. Starzyk, J.A., Liu, Y., He, H., Challenges of Embodied Intelligence, Proc. Int. Conf. on Signals and Electronic Systems, ICSES'06, Lodz, Poland, Sep. 17-20 (2006) 2. Kandel, E. R., Schwartz, J. H., Jessell, T. M., Principles of Neural Science, McGraw-Hill Medical; 4th edition (2000) 3. Anderson, J., Learning in Sparsely Connected and Sparsely Coded System, Ersatz Brain Project working note, (2005) 7

4. Barlow, H. B., Single Units and Sensation: A Neuron Doctrine for Perceptual Psychology? Perception, Vol.1, (1972) 371-394 5. Amari, S, Neural Representation of Information by Sparse Encoding, Brain Mechanisms of Perception and Memory from Neuron to Behavior, Oxford University Press, (1993) 630-637 6. Földiak, P., Young, M. P., Sparse Coding in the Primate Cortex, The Handbook of Brain Theory and Neural Networks, The MIT Press, (1995) 895-898 7. Olshausen, B. A., Field, D. J., Sparse coding of sensor inputs, Current Opinions in Neurobiology, Vol.14, (2004) 481-487 8. Haykin, S., Neural Networks: A Comprehensive Foundation, Prentice Hall (1999) 9. Zurada, J. M., Introduction to Artificial Neural Systems, West Publishing Company (1992) 10. Vogel, D. D., Boos, W., Sparsely connected, Hebbian networks with strikingly large storage capacities, Neural Networks, Vol. 10, No. 4 (1997) 671-682 11. D. D. Vogel, A neural network model of memory and higher cognitive functions, Int J Psychophysiol, Vol. 55 (1) (2005) 3-21 12. Starzyk, J. A. Fang, X., A CMOS Current Mode Winner-Take-All Circuit with both Excitatory and Inhibitory Feedback, Electronics Letters, Vol.29 (10) (1993) 908-910 13. LeCun, Y., Cortes, C., The MNIST database of handwritten digits, http://yann.lecun.com/exdb/mnist/