University of Cambridge Engineering Part IB Information Engineering Elective

Similar documents
Chapter 1. Introduction

Computational Cognitive Neuroscience

Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence

Question 1 Multiple Choice (8 marks)

Artificial Neural Networks (Ref: Negnevitsky, M. Artificial Intelligence, Chapter 6)

1. Introduction 1.1. About the content

1. Introduction 1.1. About the content. 1.2 On the origin and development of neurocomputing

Brief History of Work in the area of Learning and Memory

Sparse Coding in Sparse Winner Networks

International Journal of Scientific & Engineering Research Volume 4, Issue 2, February ISSN THINKING CIRCUIT

CHAPTER I From Biological to Artificial Neuron Model

'Automated dermatologist' detects skin cancer with expert accuracy - CNN.com

Neural Information Processing: Introduction

LESSON 3.3 WORKBOOK. Why does applying pressure relieve pain? Workbook. Postsynaptic potentials

Intelligent Control Systems

Computational Neuroscience. Instructor: Odelia Schwartz

Lecture 2.1 What is Perception?

Learning and Adaptive Behavior, Part II

LESSON 3.3 WORKBOOK. Why does applying pressure relieve pain?

Course Introduction. Neural Information Processing: Introduction. Notes. Administration

AU B. Sc.(Hon's) (Fifth Semester) Esamination, Introduction to Artificial Neural Networks-IV. Paper : -PCSC-504

Applied Neuroscience. Conclusion of Science Honors Program Spring 2017

Introduction to Computational Neuroscience

EEL-5840 Elements of {Artificial} Machine Intelligence

CS 453X: Class 18. Jacob Whitehill

Information Processing During Transient Responses in the Crayfish Visual System

Task 1: Machine Learning with Spike-Timing-Dependent Plasticity (STDP)

Neural circuits PSY 310 Greg Francis. Lecture 05. Rods and cones

VHDL implementation of Neuron based classification for future artificial intelligence applications

The Re(de)fined Neuron

A Fuzzy Improved Neural based Soft Computing Approach for Pest Disease Prediction

Processing of Logical Functions in the Human Brain

A HMM-based Pre-training Approach for Sequential Data

Introduction to Computational Neuroscience

A Neural-Fuzzy Description of Ambiguous Figures in Visual Message

Cognitive & Linguistic Sciences. What is cognitive science anyway? Why is it interdisciplinary? Why do we need to learn about information processors?

Psychology as a Science of Design in Engineering

Edge Detection Techniques Based On Soft Computing

7.1 Grading Diabetic Retinopathy

Basics of Computational Neuroscience

Oxford Foundation for Theoretical Neuroscience and Artificial Intelligence

The synaptic Basis for Learning and Memory: a Theoretical approach

Accurate Prediction of Heart Disease Diagnosing Using Computation Method

Modeling a Neuron. 1 Introduction. 2 Grade Levels and Topics. 3 Objectives. 4 Materials

Application of Artificial Neural Networks in Classification of Autism Diagnosis Based on Gene Expression Signatures

A Connectionist Model based on Physiological Properties of the Neuron

Evidence Based Diagnosis of Mesothelioma

Neuromorphic computing

Rolls,E.T. (2016) Cerebral Cortex: Principles of Operation. Oxford University Press.

TIME SERIES MODELING USING ARTIFICIAL NEURAL NETWORKS 1 P.Ram Kumar, 2 M.V.Ramana Murthy, 3 D.Eashwar, 4 M.Venkatdas

IIE 269: Cognitive Psychology

Introduction. Chapter The Perceptual Process

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012

NEOCORTICAL CIRCUITS. specifications

The issue (1) Some spikes. The issue (2) Am I spikes over time?

Implantable Microelectronic Devices

Modeling of Hippocampal Behavior

EECS 433 Statistical Pattern Recognition

BLADDERSCAN PRIME PLUS TM DEEP LEARNING

Appendix I Teaching outcomes of the degree programme (art. 1.3)

Mental Processes -- How the Mind Arises from the Brain. Roger Ellman

Lesson One: Introduction to Neuroscience Center for Sensorimotor Neural Engineering Lesson Plan Author: Phelana Pang LESSON OVERVIEW

Prediction of Malignant and Benign Tumor using Machine Learning

Embodied Cognition International Lecture Serie

A micropower support vector machine based seizure detection architecture for embedded medical devices

Lesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS

-Ensherah Mokheemer. -Amani Nofal. -Loai Alzghoul

Artificial Neural Network : Introduction

ERA: Architectures for Inference

Improving rapid counter terrorism decision making

Bottom-Up Model of Strategy Selection

An Edge-Device for Accurate Seizure Detection in the IoT

Computational Intelligence Lecture 21: Integrating Fuzzy Systems and Neural Networks

Efficacy of the Extended Principal Orthogonal Decomposition Method on DNA Microarray Data in Cancer Detection

Neural Coding. Computing and the Brain. How Is Information Coded in Networks of Spiking Neurons?

A framework for the Recognition of Human Emotion using Soft Computing models

ARTIFICIAL NEURAL NETWORKS AND THEIR APPLICATIONS IN BUSINESS

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention

Interpreting Deep Neural Networks and their Predictions

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

CSE 258 Lecture 2. Web Mining and Recommender Systems. Supervised learning Regression

MRI Image Processing Operations for Brain Tumor Detection

CAN T WE ALL JUST GET ALONG?

Evolutionary Programming

A Learning Method of Directly Optimizing Classifier Performance at Local Operating Range

Modeling a Neuron. Grade Levels. Introduction. Objectives. Materials and Resources

Neural Network Diagnosis of Avascular Necrosis from Magnetic Resonance Images

Artificial Intelligence: Its Scope and Limits, by James Fetzer, Kluver Academic Publishers, Dordrecht, Boston, London. Artificial Intelligence (AI)

ECG Beat Recognition using Principal Components Analysis and Artificial Neural Network

Diagnosing and curing diseases using Chemical Intents and Machine Learning

arxiv: v1 [cs.ai] 28 Nov 2017

Hypothesis-Driven Research

Outline. What s inside this paper? My expectation. Software Defect Prediction. Traditional Method. What s inside this paper?

Multilayer Perceptron Neural Network Classification of Malignant Breast. Mass

Development of novel algorithm by combining Wavelet based Enhanced Canny edge Detection and Adaptive Filtering Method for Human Emotion Recognition

COURSE DESCRIPTIONS 科目簡介

Outline. Animals: Nervous system. Neuron and connection of neurons. Key Concepts:

Fundamentals of Computational Neuroscience 2e

Transcription:

University of Cambridge Engineering Part IB Information Engineering Elective Paper 8: Image Searching and Modelling Using Machine Learning Handout 1: Introduction to Artificial Neural Networks Roberto Cipolla and Matthew Johnson May 17

Introduction to Neural Networks 1 Notebook You can access an interactive version of this handout at the following URL: https://aka.ms/ucepibieep8

Introduction to Neural Networks Artificial Intelligence Complex, interconnected networks of neurons, like the human brain, are the only known way of achieving general intelligence. Artificial versions of these networks have proven helpful for more targeted AI tasks, like image understanding: http://captionbot.ai An example of this can be seen in auto-captioning systems like the one above. Not only can it recognise thousands of object categories in images, but it can perform sentiment analysis on faces and produce readable, human-like captions. Try it out now! There are many publically available services like this being used to analyse images for fun or profit, including: http://what-dog.net/ http://how-old.net/ http://clarifai.com/ Systems like these can seem intelligent because they do things we usually associate with human intelligence, like scene understanding and object recognition. However, in reality, they are highly specialised systems trained from vast corpora of labeled data and constructed from elements that mimic in useful but incomplete ways the kinds of biological neural networks seen in animals. In the next three lectures we will explore the underlying mathematics behind these systems, analyse their structures and design principles, and study state of the art systems and techniques.

Introduction to Neural Networks 3 Neurons Before we dive into the math, let us perform an (extremely brief) overview of neurons and how they work. Below is a diagram of a synapse: The ways in which neurons interact involve complex interactions between electrical signals and neurochemistry that are very outside the scope of this course. That said, for our purposes it is important to understand three key concepts: 1. A neuron has both dendrites (inputs) and axons (outputs) which connect it to other neurons. A neuron will produce an exciting or inhibiting signal along its axons based upon activations from its dendrites in a non-linear way 3. Each dendrite contributes to whether a signal is produced in different proportions, which change over time We will focus on these three principles and how to use them to build a simple mathematical model of a neuron.

Introduction to Neural Networks Perceptrons The perceptron algorithm was invented in 1957 at the Cornell Aeronautical Laboratory by Frank Rosenblatt [3]. It was designed to be implemented in hardware for the purpose of image recognition, and was the marvel of the age. The New York Times reported the perceptron to be the embryo of an electronic computer that [the Navy] expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence. []. Frank Rosenblatt The Mark 1 Perceptron Rosenblatt based his designs upon earlier theoretical work by Warren McCulloch and Walter Pitts in 193 on the Threshold Logic Unit [1], arguably the first artificial neuron. The idea in both cases was to create a simple mathematical model of a neuron. In the case of the perceptron, this model was then expressed in hardware. The entire machine represented a single neuron in this sense, with inputs from a x image sensor and a binary output. It was later shown that its capabilities were fairly limited, restricted mostly to linearly separable classification problems.

Introduction to Neural Networks 5 Perceptrons, cont. The math behind perceptrons is fairly straightforward. Given an input x, the output of a perceptron is calculated as: y(x) = f(w T x) (1) where w are the weights on the input vector x and f is a step function of the form: { +1 x f(x) = () 1 x < The fundamental problem at hand is how to learn the values of w. For this we use the perceptron criterion: E P (w) = n M w T x n t n (3) where t n { 1, +1} for negative and positive examples and M is the set of misclassified examples. We can use the gradient of this function to update the weights on a per-example basis: w τ+1 = w τ η E P (w) = w τ + ηx n t n () where η is the learning rate parameter and τ is an integer that indexes the steps of the algorithm.

Introduction to Neural Networks Training a Perceptron Below is a sample problem. There are two classes, represented here by two sets of points on a plane. The perceptron s decision boundary is represented by the green line. The top left corner is the initial state. First, a red circle is presented, and the perceptron adapts its boundary. This one example solves most of this problem, and it is only when the circled blue cross in the bottom left is presented that the boundary changes again. Finally, the algorithm finds one more error in the bottom right and uses it to correct the perceptron, resulting in a perfect linear boundary between the two classes.

Introduction to Neural Networks 7 Recognising Digits An important task in the history of artificial neural networks is the recognition of handwritten digits, driven by a dataset known as MNIST. The images above were gathered from employees of the US Census Bureau and Secondary School students, consisting of approximately 5 individuals. There are, training images and 1, test images (sampled in a disjoint manner from the two sample groups). The images themselves are 8 8 unregistered grayscale images. Research has been performed on the dataset since the 198s and has continued to the present day where it is often used in tutorials for neural net software frameworks. In Rosenblatt s original work, he trained perceptrons to tell the difference between two different images of numbers. Using this new dataset, we can replicate those experiments, though we will do it entirely in software of course.

Introduction to Neural Networks 8 1 vs First, we look at the result of training a perceptron to distinguish between hand-written 1s and s from the MNIST dataset. Below are snapshots from the nd, 717th and th image presented to the perceptron: 1..75.5.5..5.5.75 1. 1..8.... 5 1 15 3 1 1 1..8.... 5 1 15 5 3 1 1 3 1..8.... 5 1 15 As you can see, the perceptron is quite effective at solving this problem, achieving a final accuracy of.998. It clearly learns that the positive class (1) has a vertical bit in the center, while the negative class () is shaped like a ring.

Introduction to Neural Networks 9 vs 5 Now we will examine what happens when we train a perceptron to distinguish two quite similar numerals, and 5. Below are snapshots from the nd, 55th and th image presented to the perceptron: 1..75.5.5..5.5.75 1. 1..8.... 5 1 15 1..8.... 5 1 15 7.5 5..5..5 5. 7.5 1. 1..8.... 5 1 15 The perceptron seems to have more difficulty with this task, achieving a lower accuracy of.958 and seeming to plateau. Indeed, if this perceptron is presented with more examples it will never converge and eventually causes a numerical overflow.

Introduction to Neural Networks 1 Visualisation Why does the perceptron solve one problem almost perfectly, and seem to stumble on the second? Below are some helpful visualizations of a low-dimensional embedding of the different pairs of image classes. 8 Each point represents a different digit image. Note how, even in this low-dimensional space, the 1s and s are linearly separable? Contrast this with the s and 5s, which are all mixed up. There is not a line that separates these two classes, and that is why the perceptron will never perfectly achieve the task of distinguishing them. In the next lecture, we will examine how to combine perceptrons in different ways to overcome these limitations.

Introduction to Neural Networks 11 References [1] Warren S. McCulloch and Walter Pitts. A logical calculus of the ideas immanent in nervous activity. Bulletin of Mathematical Biophysics, 5:115 133, 193. [] Mikel Olazaran. A sociological study of the official history of the perceptrons controversy. Social Studies of Science, (3):11 59, 199. [3] Frank Rosenblatt. The Perceptron - a perceiving and recognizing automaton. Technical Report 85--1, Cornell Aeronautical Laboratory, 1957.