Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6)
|
|
- Lesley Watkins
- 5 years ago
- Views:
Transcription
1 Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6) BPNN in Practice Week 3 Lecture Notes page 1 of 1
2 The Hopfield Network In this network, it was designed on analogy of brain s memory, which is work by association. For example, we can recognise a familiar face even in an unfamiliar environment within ms. We can also recall a complete sensory experience, including sounds and scenes, when we hear only a few bars of music. The brain routinely associates one thing with another. Multilayer Neural Networks trained with backpropagation algorithm are used for pattern recognition problems. To emulate the human memory s associative characteristics, we use a recurrent neural network. Week 3 Lecture Notes page 2 of 2
3 A recurrent neural network has feedback loops from its outputs to its inputs. The presence of such loops has a profound impact on the learning capability of the network. Single layer n-neuron Hopfield network The Hopfield network uses McCulloch and Pitts neurons with the sign activation function as its computing element: The current state of the Hopfield is determined by the current outputs of all neurons, y 1, y 2,,y n. Week 3 Lecture Notes page 3 of 3
4 Hopfield Learning Algorithm: Step 1: Assign weights o Assign random connections weights with values w = +1 or w = 1 for all i j and ij 0 for i = j ij Step 2: Initialisation o Initialise the network with an unknown pattern: x i = whereo and x + 1or -1 O ( k), 0 i i i i ( k) is N 1, theoutput of is an element at input i of nodei at time t = k = an input pattern, 0 Step 3: Convergence o Iterate until convergence is reached, using the relation: O i ( k N 1 + 1) = f i= 0 w ij O i ( k), 0 j N 1 where the function f(.) is a hard limiting nonlinearity. o Repeat the process until the node outputs remain unchanged. Week 3 Lecture Notes page 4 of 4
5 o The node outputs then best represent the exemplar pattern that best matches the unknown input. Step 4: Repeat for Next Pattern o Go back to step 2 and repeat for next x i, and so on. Hopfield network can act as an error correction network. Type of Learning Supervised Learning o the input vectors and the corresponding output vectors are given o the ANN learns to approximate the function from the inputs to the outputs Week 3 Lecture Notes page 5 of 5
6 Reinforcement Learning o the input vectors and a reinforcement signal are given o the reinforcement signal tells how good the true output was Unsupervised Learning o only input are given o the ANN learns to form internal representations or codes for the input data that can then be used e.g. for clustering From now we will look at unsupervised learning neural networks. Week 3 Lecture Notes page 6 of 6
7 Hebbian Learning In 1949, Donald Hebb proposed one of the key ideas in biological learning, commonly known as Hebb s Law. Hebb s Law states that if neuron i is near enough to excite neuron j and repeatedly participates in its activation, the synaptic connection between these two neurons is strengthened and neuron j becomes more sensitive to stimuli from neuron i. Hebb s Law can be represented in the form of two rules: If two neurons on either side of a connection are activated synchronously, then the weight of that connection is increased. If two neurons on either side of a connection are activated asynchronously, then the weight of that connection is decreased. Week 3 Lecture Notes page 7 of 7
8 Hebbian learning implies that weights can only increase. To resolve this problem, we might impose a limit on the growth of synaptic weights. It can be implemented by introducing a non-linear forgetting factor into Hebb s Law: where ö is the following factor Forgetting factor usually falls in the interval between 0 and 1, typically between 0.01 and 0.1, to allow only a little forgetting while limiting the weight growth. Week 3 Lecture Notes page 8 of 8
9 Hebbian Learning Algorithm: Step 1: Initialisation o Set initial synaptic weights and threshold to small random values, say in an interval [0,1]. Step 2: Activation o Compute the neuron output at iteraction p where n is the number of neuron inputs, and è j is the threshold value of neuron j. Step 3: Learning o Update the weights in the network: where Äw ij (p) is the weight correction at iteration p. Week 3 Lecture Notes page 9 of 9
10 o The weight correction is determine by the generalised activity product rule: Step 4: Iteration Increase iteration p by one, go back to Step 2. Competitive Learning Neurons compete among themselves to be activated While in Hebbian Learning, several output neurons can be activated simultaneously, in competitive learning, only a single output neuron is active at any time. The output neuron that wins the competition is called the winner-takes-all neuron. In the late 1980s, Kohonen introduced a special call of ANN called self-organising maps. These maps are based on competitive learning. Week 3 Lecture Notes page 10 of 10
11 Self-organising Map Our brain is dominated by the cerebral cortex, a very complex structure of billions of neurons and hundreds of billions of synapses. The cortex includes areas that are responsible for different human activities (motor, visual, auditory, etc), and associated with different sensory input. Each sensory input is mapped into a corresponding area of the cerebral cortex. The cortex is a selforganising computational map in the human brain. Week 3 Lecture Notes page 11 of 11
12 Feature-mapping Kohonen model The Kohonen Network The Kohonen model provides a topological mapping. It places a fixed number of input patterns from the input layer into a higher dimensional output or Kohonen layer. Week 3 Lecture Notes page 12 of 12
13 Training in the Kohonen network begins with the winner s neighbourhood of a fairly large size. Then, as training proceeds, the neighbourhood size gradually decreases. The lateral connections are used to create a competition between neurons. The neuron with the largest activation level among all neurons in the output layer becomes the winner. The winning neuron is the only neuron that produces an output signal. The activity of all other neurons is suppressed in the competition. Week 3 Lecture Notes page 13 of 13
14 The lateral feedback connections produce excitatory or inhibitory effects, depending on the distance from the winning neuron. This is achieved by the use of a Mexican hat function which describes synaptic weights between neurons in the Kohonen layer. In the Kohonen network, a neuron learns by shifting its weights from inactive connections to actives ones. Only the winning neuron and its neighbourhood are allowed to learn. Week 3 Lecture Notes page 14 of 14
15 Competitive Learning Algorithm Step 1: Initialisation o Set initial synaptic weights to small random values, say in an interval [0, 1], and assign a small positive value to learning rate parameter á Step 2: Activation and Similarity Matching o Activate the Kohonen network by applying the input vector X, and find the winner-takes-all (best matching) neuron j X at iteration p, using the minimum-distance Euclidean criterion where n is the number of neurons in the input layer, and m is the number of neurons in the Kohonen layer. Week 3 Lecture Notes page 15 of 15
16 Step 3: Learning o Update the synaptic weights where Äw ij (p) is the weight correction at iteration p. o The weight correction is determined by the competitive learning rule: where á is the learning rate parameter, and Ë j (p) is the neighbourhood function central around the winner-takes-all neuron j x at iteration p. Step 4: Iteration o Increase iteration p by one, go back to step 2 and continue until minimum-distance Euclidean criterion is satisfied, or no noticeable changes occur in the feature map Week 3 Lecture Notes page 16 of 16
Learning in neural networks
http://ccnl.psy.unipd.it Learning in neural networks Marco Zorzi University of Padova M. Zorzi - European Diploma in Cognitive and Brain Sciences, Cognitive modeling", HWK 19-24/3/2006 1 Connectionist
More informationCognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence
Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence To understand the network paradigm also requires examining the history
More informationThe storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls
The storage and recall of memories in the hippocampo-cortical system Supplementary material Edmund T Rolls Oxford Centre for Computational Neuroscience, Oxford, England and University of Warwick, Department
More informationModeling of Hippocampal Behavior
Modeling of Hippocampal Behavior Diana Ponce-Morado, Venmathi Gunasekaran and Varsha Vijayan Abstract The hippocampus is identified as an important structure in the cerebral cortex of mammals for forming
More informationCell Responses in V4 Sparse Distributed Representation
Part 4B: Real Neurons Functions of Layers Input layer 4 from sensation or other areas 3. Neocortical Dynamics Hidden layers 2 & 3 Output layers 5 & 6 to motor systems or other areas 1 2 Hierarchical Categorical
More information1. Introduction 1.1. About the content
1. Introduction 1.1. About the content At first, some background ideas are given and what the origins of neurocomputing and artificial neural networks were. Then we start from single neurons or computing
More information1. Introduction 1.1. About the content. 1.2 On the origin and development of neurocomputing
1. Introduction 1.1. About the content At first, some background ideas are given and what the origins of neurocomputing and artificial neural networks were. Then we start from single neurons or computing
More informationSparse Coding in Sparse Winner Networks
Sparse Coding in Sparse Winner Networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University, Athens, OH 45701 {starzyk, yliu}@bobcat.ent.ohiou.edu
More informationExploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells
Neurocomputing, 5-5:389 95, 003. Exploring the Functional Significance of Dendritic Inhibition In Cortical Pyramidal Cells M. W. Spratling and M. H. Johnson Centre for Brain and Cognitive Development,
More informationRolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.
Digital Signal Processing and the Brain Is the brain a digital signal processor? Digital vs continuous signals Digital signals involve streams of binary encoded numbers The brain uses digital, all or none,
More informationLearning and Adaptive Behavior, Part II
Learning and Adaptive Behavior, Part II April 12, 2007 The man who sets out to carry a cat by its tail learns something that will always be useful and which will never grow dim or doubtful. -- Mark Twain
More informationComputational Cognitive Neuroscience
Computational Cognitive Neuroscience Computational Cognitive Neuroscience Computational Cognitive Neuroscience *Computer vision, *Pattern recognition, *Classification, *Picking the relevant information
More informationNeural Networks. Nice books to start reading:
Neural Networks Overview: - Anatomy of Neuronal Networks - Formal Neural Networks - Are they realistic? - Oscillations and Phase locking - Mapping problem: Kohonen Networks Nice books to start reading:
More informationIntelligent Control Systems
Lecture Notes in 4 th Class in the Control and Systems Engineering Department University of Technology CCE-CN432 Edited By: Dr. Mohammed Y. Hassan, Ph. D. Fourth Year. CCE-CN432 Syllabus Theoretical: 2
More informationQuestion 1 Multiple Choice (8 marks)
Philadelphia University Student Name: Faculty of Engineering Student Number: Dept. of Computer Engineering First Exam, First Semester: 2015/2016 Course Title: Neural Networks and Fuzzy Logic Date: 19/11/2015
More informationA Model of Visually Guided Plasticity of the Auditory Spatial Map in the Barn Owl
A Model of Visually Guided Plasticity of the Auditory Spatial Map in the Barn Owl Andrea Haessly andrea@cs.utexas.edu Joseph Sirosh sirosh@cs.utexas.edu Risto Miikkulainen risto@cs.utexas.edu Abstract
More informationIntroduction to Computational Neuroscience
Introduction to Computational Neuroscience Lecture 7: Network models Lesson Title 1 Introduction 2 Structure and Function of the NS 3 Windows to the Brain 4 Data analysis 5 Data analysis II 6 Single neuron
More informationArtificial Neural Networks
Artificial Neural Networks Torsten Reil torsten.reil@zoo.ox.ac.uk Outline What are Neural Networks? Biological Neural Networks ANN The basics Feed forward net Training Example Voice recognition Applications
More informationLesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS
Lesson 6 Learning II Anders Lyhne Christensen, D6.05, anders.christensen@iscte.pt INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS First: Quick Background in Neural Nets Some of earliest work in neural networks
More informationPlasticity of Cerebral Cortex in Development
Plasticity of Cerebral Cortex in Development Jessica R. Newton and Mriganka Sur Department of Brain & Cognitive Sciences Picower Center for Learning & Memory Massachusetts Institute of Technology Cambridge,
More informationAbstract A neural network model called LISSOM for the cooperative self-organization of
Modeling Cortical Plasticity Based on Adapting Lateral Interaction Joseph Sirosh and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin, Austin, TX{78712. email: sirosh,risto@cs.utexas.edu
More informationUniversity of Cambridge Engineering Part IB Information Engineering Elective
University of Cambridge Engineering Part IB Information Engineering Elective Paper 8: Image Searching and Modelling Using Machine Learning Handout 1: Introduction to Artificial Neural Networks Roberto
More informationAU B. Sc.(Hon's) (Fifth Semester) Esamination, Introduction to Artificial Neural Networks-IV. Paper : -PCSC-504
AU-6919 B. Sc.(Hon's) (Fifth Semester) Esamination, 2014 Introduction to Artificial Neural Networks-IV Paper : -PCSC-504 Section-A 1. (I) Synaptic Weights provides the... method for the design of neural
More informationMemory, Attention, and Decision-Making
Memory, Attention, and Decision-Making A Unifying Computational Neuroscience Approach Edmund T. Rolls University of Oxford Department of Experimental Psychology Oxford England OXFORD UNIVERSITY PRESS Contents
More informationCerebral Cortex. Edmund T. Rolls. Principles of Operation. Presubiculum. Subiculum F S D. Neocortex. PHG & Perirhinal. CA1 Fornix CA3 S D
Cerebral Cortex Principles of Operation Edmund T. Rolls F S D Neocortex S D PHG & Perirhinal 2 3 5 pp Ento rhinal DG Subiculum Presubiculum mf CA3 CA1 Fornix Appendix 4 Simulation software for neuronal
More informationCHAPTER I From Biological to Artificial Neuron Model
CHAPTER I From Biological to Artificial Neuron Model EE543 - ANN - CHAPTER 1 1 What you see in the picture? EE543 - ANN - CHAPTER 1 2 Is there any conventional computer at present with the capability of
More informationHierarchical dynamical models of motor function
ARTICLE IN PRESS Neurocomputing 70 (7) 975 990 www.elsevier.com/locate/neucom Hierarchical dynamical models of motor function S.M. Stringer, E.T. Rolls Department of Experimental Psychology, Centre for
More informationNeurons and neural networks II. Hopfield network
Neurons and neural networks II. Hopfield network 1 Perceptron recap key ingredient: adaptivity of the system unsupervised vs supervised learning architecture for discrimination: single neuron perceptron
More informationCAS Seminar - Spiking Neurons Network (SNN) Jakob Kemi ( )
CAS Seminar - Spiking Neurons Network (SNN) Jakob Kemi (820622-0033) kemiolof@student.chalmers.se November 20, 2006 Introduction Biological background To be written, lots of good sources. Background First
More informationSpiking Inputs to a Winner-take-all Network
Spiking Inputs to a Winner-take-all Network Matthias Oster and Shih-Chii Liu Institute of Neuroinformatics University of Zurich and ETH Zurich Winterthurerstrasse 9 CH-857 Zurich, Switzerland {mao,shih}@ini.phys.ethz.ch
More informationlateral connections are non-modiable and distinct from with short-range excitation and long-range inhibition.
How Lateral Interaction Develops in a Self-Organizing Feature Map Joseph Sirosh and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin, Austin, TX-771 sirosh,risto@cs.utexas.edu
More informationRealization of Visual Representation Task on a Humanoid Robot
Istanbul Technical University, Robot Intelligence Course Realization of Visual Representation Task on a Humanoid Robot Emeç Erçelik May 31, 2016 1 Introduction It is thought that human brain uses a distributed
More informationUniversity of California Postprints
University of California Postprints Year 2006 Paper 2444 Analysis of early hypoxia EEG based on a novel chaotic neural network M Hu G Li J J. Li Walter J. Freeman University of California, Berkeley M Hu,
More informationPMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos
PMR5406 Redes Neurais e Lógica Fuzzy Aula 5 Alguns Exemplos APPLICATIONS Two examples of real life applications of neural networks for pattern classification: RBF networks for face recognition FF networks
More informationTIME SERIES MODELING USING ARTIFICIAL NEURAL NETWORKS 1 P.Ram Kumar, 2 M.V.Ramana Murthy, 3 D.Eashwar, 4 M.Venkatdas
TIME SERIES MODELING USING ARTIFICIAL NEURAL NETWORKS 1 P.Ram Kumar, 2 M.V.Ramana Murthy, 3 D.Eashwar, 4 M.Venkatdas 1 Department of Computer Science & Engineering,UCE,OU,Hyderabad 2 Department of Mathematics,UCS,OU,Hyderabad
More informationThe Ever-Changing Brain. Dr. Julie Haas Biological Sciences
The Ever-Changing Brain Dr. Julie Haas Biological Sciences Outline 1) Synapses: excitatory, inhibitory, and gap-junctional 2) Synaptic plasticity, and Hebb s postulate 3) Sensory maps and plasticity 4)
More informationAssociative Memory-I: Storing Patterns
Printed from the Mathematica Help Browser 1 «1 of 11 Associative Memory-I: Storing Patterns Learning and recalling memories are important to animals' survival. Navigational memory is important for locating
More informationClusters, Symbols and Cortical Topography
Clusters, Symbols and Cortical Topography Lee Newman Thad Polk Dept. of Psychology Dept. Electrical Engineering & Computer Science University of Michigan 26th Soar Workshop May 26, 2006 Ann Arbor, MI agenda
More informationChapter 1. Introduction
Chapter 1 Introduction Artificial neural networks are mathematical inventions inspired by observations made in the study of biological systems, though loosely based on the actual biology. An artificial
More informationSynaptic theory of gradient learning with empiric inputs. Ila Fiete Kavli Institute for Theoretical Physics. Types of learning
Synaptic theory of gradient learning with empiric inputs Ila Fiete Kavli Institute for Theoretical Physics Types of learning Associations among events (and reinforcements). e.g., classical conditioning
More informationThe following article was presented as an oral presentation at the Conference ICANN 98:
The following article was presented as an oral presentation at the Conference ICANN 98: Wiemer J, Spengler F, Joublin F, Stagge P, Wacquant S: A Model of Cortical Plasticity: Integration and Segregation
More informationActive Sites model for the B-Matrix Approach
Active Sites model for the B-Matrix Approach Krishna Chaithanya Lingashetty Abstract : This paper continues on the work of the B-Matrix approach in hebbian learning proposed by Dr. Kak. It reports the
More informationA Comparison of Supervised Multilayer Back Propagation and Unsupervised Self Organizing Maps for the Diagnosis of Thyroid Disease
A Comparison of Supervised Multilayer Back Propagation and Unsupervised Self Organizing Maps for the Diagnosis of Thyroid Disease Arpneek Kaur M. Tech Student CT Instt. of Engg., Mgt. & Narpat Singh M.
More informationLecture 1: Neurons. Lecture 2: Coding with spikes. To gain a basic understanding of spike based neural codes
Lecture : Neurons Lecture 2: Coding with spikes Learning objectives: To gain a basic understanding of spike based neural codes McCulloch Pitts Neuron I w in Σ out Θ Examples: I = ; θ =.5; w=. - in = *.
More informationProf. Greg Francis 7/31/15
s PSY 200 Greg Francis Lecture 06 How do you recognize your grandmother? Action potential With enough excitatory input, a cell produces an action potential that sends a signal down its axon to other cells
More informationWhy do we have a hippocampus? Short-term memory and consolidation
Why do we have a hippocampus? Short-term memory and consolidation So far we have talked about the hippocampus and: -coding of spatial locations in rats -declarative (explicit) memory -experimental evidence
More informationObservational Learning Based on Models of Overlapping Pathways
Observational Learning Based on Models of Overlapping Pathways Emmanouil Hourdakis and Panos Trahanias Institute of Computer Science, Foundation for Research and Technology Hellas (FORTH) Science and Technology
More informationNeural Network Modelling of Autism
Neural Network Modelling of Autism Recent developments in autism research (M. F. Casanova, ed.), pp. 100 134, Hauppauge, New York: Nova Science Publishers, Inc., 2005. In press. Lennart Gustafsson, Luleå
More informationA toy model of the brain
A toy model of the brain arxiv:q-bio/0405002v [q-bio.nc] 2 May 2004 B. Hoeneisen and F. Pasmay Universidad San Francisco de Quito 30 March 2004 Abstract We have designed a toy brain and have written computer
More informationContinuous transformation learning of translation invariant representations
Exp Brain Res (21) 24:255 27 DOI 1.17/s221-1-239- RESEARCH ARTICLE Continuous transformation learning of translation invariant representations G. Perry E. T. Rolls S. M. Stringer Received: 4 February 29
More informationSynfire chains with conductance-based neurons: internal timing and coordination with timed input
Neurocomputing 5 (5) 9 5 www.elsevier.com/locate/neucom Synfire chains with conductance-based neurons: internal timing and coordination with timed input Friedrich T. Sommer a,, Thomas Wennekers b a Redwood
More informationUMIACS-TR August Feature Maps. Yinong Chen. University of Maryland. College Park, MD 20742
UMIACS-TR-97-56 August 1997 CS-TR-3816 A Motor Control Model Based on Self-organizing Feature Maps Yinong Chen Institute for Advanced Computer Studies Department of Computer Science University of Maryland
More informationInvestigation of Physiological Mechanism For Linking Field Synapses
Investigation of Physiological Mechanism For Linking Field Synapses Richard B. Wells 1, Nick Garrett 2, Tom Richner 3 Microelectronics Research and Communications Institute (MRCI) BEL 316 University of
More informationVISUAL CORTICAL PLASTICITY
VISUAL CORTICAL PLASTICITY OCULAR DOMINANCE AND OTHER VARIETIES REVIEW OF HIPPOCAMPAL LTP 1 when an axon of cell A is near enough to excite a cell B and repeatedly and consistently takes part in firing
More informationHow has Computational Neuroscience been useful? Virginia R. de Sa Department of Cognitive Science UCSD
How has Computational Neuroscience been useful? 1 Virginia R. de Sa Department of Cognitive Science UCSD What is considered Computational Neuroscience? 2 What is considered Computational Neuroscience?
More informationIntroduction. Chapter The Perceptual Process
Chapter 1 Introduction Most of us take for granted our ability to perceive the external world. However, this is no simple deed at all. Imagine being given a task of designing a machine that can perceive,
More informationarxiv:adap-org/ v1 28 Jul 1997
Learning from mistakes Dante R. Chialvo* and Per Bak* Niels Bohr Institute, Blegdamsvej 17, Copenhagen, Denmark. Division of Neural Systems, Memory and Aging, University of Arizona, Tucson, AZ 85724, USA.
More informationRetinal Waves and Ocular Dominance Columns
Retinal Waves and Ocular Dominance Columns In the cat, at birth, inputs from both eyes are intermingled in the visual cortex. Ocular dominance columns start to appear a few weeks after birth. They can
More informationA Stochastic and Competitive Hebbian Learning Mechanism through Spike-Timing Dependent Plasticity to Lift Hard Weight Constraints on Hebbian Synapses
Nagaoka University of Technology Graduate School of Engineering Information Science and Control Engineering A Stochastic and Competitive Hebbian Learning Mechanism through Spike-Timing Dependent Plasticity
More informationYesterday s Picture UNIT 3F
Warm-Up Create a representation (cartoon, diagram, flow map, et cetera) to show how the vertebrate nervous system detects transmits information from one neuron to another. (LO 3.49) Yesterday s Picture
More informationNew Ideas for Brain Modelling
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 06, Issue10 (Oct. 2016), V1 PP 13-17 www.iosrjen.org New Ideas for Brain Modelling Kieran Greer 1 1 Distributed Computing
More informationAnalysis of Mammograms Using Texture Segmentation
Analysis of Mammograms Using Texture Segmentation Joel Quintanilla-Domínguez 1, Jose Miguel Barrón-Adame 1, Jose Antonio Gordillo-Sosa 1, Jose Merced Lozano-Garcia 2, Hector Estrada-García 2, Rafael Guzmán-Cabrera
More informationTheoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück
The Binding Problem This lecture is based on following articles: Adina L. Roskies: The Binding Problem; Neuron 1999 24: 7 Charles M. Gray: The Temporal Correlation Hypothesis of Visual Feature Integration:
More informationEvaluating the Effect of Spiking Network Parameters on Polychronization
Evaluating the Effect of Spiking Network Parameters on Polychronization Panagiotis Ioannou, Matthew Casey and André Grüning Department of Computing, University of Surrey, Guildford, Surrey, GU2 7XH, UK
More informationBrain anatomy and artificial intelligence. L. Andrew Coward Australian National University, Canberra, ACT 0200, Australia
Brain anatomy and artificial intelligence L. Andrew Coward Australian National University, Canberra, ACT 0200, Australia The Fourth Conference on Artificial General Intelligence August 2011 Architectures
More informationLateral Inhibition Explains Savings in Conditioning and Extinction
Lateral Inhibition Explains Savings in Conditioning and Extinction Ashish Gupta & David C. Noelle ({ashish.gupta, david.noelle}@vanderbilt.edu) Department of Electrical Engineering and Computer Science
More informationSelf-Organization and Segmentation with Laterally Connected Spiking Neurons
Self-Organization and Segmentation with Laterally Connected Spiking Neurons Yoonsuck Choe Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 USA Risto Miikkulainen Department
More informationA Neural Network Model of Naive Preference and Filial Imprinting in the Domestic Chick
A Neural Network Model of Naive Preference and Filial Imprinting in the Domestic Chick Lucy E. Hadden Department of Cognitive Science University of California, San Diego La Jolla, CA 92093 hadden@cogsci.ucsd.edu
More informationThis article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and
This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution
More informationJ Jeffress model, 3, 66ff
Index A Absolute pitch, 102 Afferent projections, inferior colliculus, 131 132 Amplitude modulation, coincidence detector, 152ff inferior colliculus, 152ff inhibition models, 156ff models, 152ff Anatomy,
More informationInternational Journal of Scientific & Engineering Research Volume 4, Issue 2, February ISSN THINKING CIRCUIT
International Journal of Scientific & Engineering Research Volume 4, Issue 2, February-2013 1 THINKING CIRCUIT Mr.Mukesh Raju Bangar Intern at Govt. Dental College and hospital, Nagpur Email: Mukeshbangar008@gmail.com
More informationPHY3111 Mid-Semester Test Study. Lecture 2: The hierarchical organisation of vision
PHY3111 Mid-Semester Test Study Lecture 2: The hierarchical organisation of vision 1. Explain what a hierarchically organised neural system is, in terms of physiological response properties of its neurones.
More information2 Physiological and Psychological Foundations
2 Physiological and Psychological Foundations 2.1 Human Nervous System 2.2 Human Brain 2.3 Human Memory 2.4 Remembering and Forgetting 1 About This Class: Flipped Classroom Source: washington.edu Four
More informationOscillatory Neural Network for Image Segmentation with Biased Competition for Attention
Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention Tapani Raiko and Harri Valpola School of Science and Technology Aalto University (formerly Helsinki University of
More informationNEOCORTICAL CIRCUITS. specifications
NEOCORTICAL CIRCUITS specifications where are we coming from? human-based computing using typically human faculties associating words with images -> labels for image search locating objects in images ->
More informationPhysiology Unit 2 CONSCIOUSNESS, THE BRAIN AND BEHAVIOR
Physiology Unit 2 CONSCIOUSNESS, THE BRAIN AND BEHAVIOR In Physiology Today What the Brain Does The nervous system determines states of consciousness and produces complex behaviors Any given neuron may
More informationASSOCIATIVE MEMORY AND HIPPOCAMPAL PLACE CELLS
International Journal of Neural Systems, Vol. 6 (Supp. 1995) 81-86 Proceedings of the Neural Networks: From Biology to High Energy Physics @ World Scientific Publishing Company ASSOCIATIVE MEMORY AND HIPPOCAMPAL
More informationPHGY 210,2,4 - Physiology SENSORY PHYSIOLOGY. Martin Paré
PHGY 210,2,4 - Physiology SENSORY PHYSIOLOGY Martin Paré Associate Professor of Physiology & Psychology pare@biomed.queensu.ca http://brain.phgy.queensu.ca/pare PHGY 210,2,4 - Physiology SENSORY PHYSIOLOGY
More informationCOGNITIVE SCIENCE 107A. Hippocampus. Jaime A. Pineda, Ph.D.
COGNITIVE SCIENCE 107A Hippocampus Jaime A. Pineda, Ph.D. Common (Distributed) Model of Memory Processes Time Course of Memory Processes Long Term Memory DECLARATIVE NON-DECLARATIVE Semantic Episodic Skills
More informationAll questions below pertain to mandatory material: all slides, and mandatory homework (if any).
ECOL 182 Spring 2008 Dr. Ferriere s lectures Lecture 6: Nervous system and brain Quiz Book reference: LIFE-The Science of Biology, 8 th Edition. http://bcs.whfreeman.com/thelifewire8e/ All questions below
More informationChapter 3: Information Processing
SENG 5334: Human Factors Engineering & INDH 5931: Research Topics in IH/Safety Chapter 3: Information Processing By: Magdy Akladios, PhD, PE, CSP, CPE, CSHM 1 A Model of Information Processing Def: A model
More informationCISC 3250 Systems Neuroscience
CISC 3250 Systems Neuroscience Levels of organization Central Nervous System 1m 10 11 neurons Neural systems and neuroanatomy Systems 10cm Networks 1mm Neurons 100μm 10 8 neurons Professor Daniel Leeds
More informationMulti-Associative Memory in flif Cell Assemblies
Multi-Associative Memory in flif Cell Assemblies Christian R. Huyck (c.huyck@mdx.ac.uk) Kailash Nadh (k.nadh@mdx.ac.uk) School of Engineering and Information Sciences Middlesex University London, UK Abstract
More informationA Neural Network Architecture for.
A Neural Network Architecture for Self-Organization of Object Understanding D. Heinke, H.-M. Gross Technical University of Ilmenau, Division of Neuroinformatics 98684 Ilmenau, Germany e-mail: dietmar@informatik.tu-ilmenau.de
More informationComputational Explorations in Cognitive Neuroscience Chapter 7: Large-Scale Brain Area Functional Organization
Computational Explorations in Cognitive Neuroscience Chapter 7: Large-Scale Brain Area Functional Organization 1 7.1 Overview This chapter aims to provide a framework for modeling cognitive phenomena based
More informationNeuromorphic computing
Neuromorphic computing Robotics M.Sc. programme in Computer Science lorenzo.vannucci@santannapisa.it April 19th, 2018 Outline 1. Introduction 2. Fundamentals of neuroscience 3. Simulating the brain 4.
More informationVS : Systemische Physiologie - Animalische Physiologie für Bioinformatiker. Neuronenmodelle III. Modelle synaptischer Kurz- und Langzeitplastizität
Bachelor Program Bioinformatics, FU Berlin VS : Systemische Physiologie - Animalische Physiologie für Bioinformatiker Synaptische Übertragung Neuronenmodelle III Modelle synaptischer Kurz- und Langzeitplastizität
More informationAnalysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014
Analysis of in-vivo extracellular recordings Ryan Morrill Bootcamp 9/10/2014 Goals for the lecture Be able to: Conceptually understand some of the analysis and jargon encountered in a typical (sensory)
More informationPSY 215 Lecture 13 (3/7/11) Learning & Memory Dr. Achtman PSY 215. Lecture 13 Topic: Mechanisms of Learning and Memory Chapter 13, section 13.
PSY 215 Lecture 13 Topic: Mechanisms of Learning and Memory Chapter 13, section 13.2 Corrections: No corrections Announcements: Question # 37 was thrown out on the last test because it was confusing the
More informationImproving Associative Memory in a Network of Spiking
Improving Associative Memory in a Network of Spiking Neurons Computing Science and Mathematics School of Natural Sciences University of Stirling Scotland FK9 4LA Thesis submitted for the degree of Doctor
More informationHebbian Plasticity for Improving Perceptual Decisions
Hebbian Plasticity for Improving Perceptual Decisions Tsung-Ren Huang Department of Psychology, National Taiwan University trhuang@ntu.edu.tw Abstract Shibata et al. reported that humans could learn to
More informationNEURAL SYSTEMS FOR INTEGRATING ROBOT BEHAVIOURS
NEURAL SYSTEMS FOR INTEGRATING ROBOT BEHAVIOURS Brett Browning & Gordon Wyeth University of Queensland Computer Science and Electrical Engineering Department Email: browning@elec.uq.edu.au & wyeth@elec.uq.edu.au
More informationThalamocortical Feedback and Coupled Oscillators
Thalamocortical Feedback and Coupled Oscillators Balaji Sriram March 23, 2009 Abstract Feedback systems are ubiquitous in neural systems and are a subject of intense theoretical and experimental analysis.
More informationModule 1 CREATE. Diagram. Getting the hardware sorted: How your brain works. Outside appearance of the human brain
CREATE Getting the hardware sorted: How your brain works Your cognition is your ability to think, learn and remember. The human brain has been described as the most complex thing in the universe. To get
More informationA Model of Perceptual Change by Domain Integration
A Model of Perceptual Change by Domain Integration Gert Westermann (gert@csl.sony.fr) Sony Computer Science Laboratory 6 rue Amyot 755 Paris, France Abstract A neural network model is presented that shows
More informationA Dynamic Field Theory of Visual Recognition in Infant Looking Tasks
A Dynamic Field Theory of Visual Recognition in Infant Looking Tasks Sammy Perone (sammy-perone@uiowa.edu) and John P. Spencer (john-spencer@uiowa.edu) Department of Psychology, 11 Seashore Hall East Iowa
More informationPosition invariant recognition in the visual system with cluttered environments
PERGAMON Neural Networks 13 (2000) 305 315 Contributed article Position invariant recognition in the visual system with cluttered environments S.M. Stringer, E.T. Rolls* Oxford University, Department of
More informationRepresenting Where along with What Information in a Model of a Cortical Patch
Representing Where along with What Information in a Model of a Cortical Patch Yasser Roudi 1,2 *, Alessandro Treves 2,3 1 Gatsby Computational Neuroscience Unit, UCL, United Kingdom, 2 Cognitive Neuroscience
More informationChapter 2--Introduction to the Physiology of Perception
Chapter 2--Introduction to the Physiology of Perception Student: 1. Our perception of the environment depends on A. the properties of the objects in the environment. B. the properties of the electrical
More information