CS 453X: Class 18. Jacob Whitehill

Similar documents
Question 1 Multiple Choice (8 marks)

Chapter 1. Introduction

Artificial Neural Network : Introduction

Intelligent Control Systems

Neurons: Structure and communication

Structure of a Neuron:

1. Introduction 1.1. About the content

1. Introduction 1.1. About the content. 1.2 On the origin and development of neurocomputing

CHAPTER I From Biological to Artificial Neuron Model

Learning in neural networks

Applied Neuroscience. Conclusion of Science Honors Program Spring 2017

Introduction to Computational Neuroscience

Artificial Neural Networks and Near Infrared Spectroscopy - A case study on protein content in whole wheat grain

Anatomy Review Graphics are used with permission of: Pearson Education Inc., publishing as Benjamin Cummings (

The storage and recall of memories in the hippocampo-cortical system. Supplementary material. Edmund T Rolls

International Journal of Scientific & Engineering Research Volume 4, Issue 2, February ISSN THINKING CIRCUIT

Learning and Adaptive Behavior, Part II

Dayalbagh Educational Institute Soft Computing Edge Cutting Technology Lab (DEISEL)

Action potential. Definition: an all-or-none change in voltage that propagates itself down the axon

COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks

CSE Introduction to High-Perfomance Deep Learning ImageNet & VGG. Jihyung Kil

TIME SERIES MODELING USING ARTIFICIAL NEURAL NETWORKS 1 P.Ram Kumar, 2 M.V.Ramana Murthy, 3 D.Eashwar, 4 M.Venkatdas

Efficacy of the Extended Principal Orthogonal Decomposition Method on DNA Microarray Data in Cancer Detection

On Training of Deep Neural Network. Lornechen

ANN predicts locoregional control using molecular marker profiles of. Head and Neck squamous cell carcinoma

Neuromorphic computing

Computational Cognitive Neuroscience

Lecture 22: A little Neurobiology

VHDL implementation of Neuron based classification for future artificial intelligence applications

Introduction to Computational Neuroscience

Cognitive Neuroscience History of Neural Networks in Artificial Intelligence The concept of neural network in artificial intelligence

BIOLOGICAL PROCESSES

Brief History of Work in the area of Learning and Memory

MRI Image Processing Operations for Brain Tumor Detection

Evidence Based Diagnosis of Mesothelioma

35-2 The Nervous System Slide 1 of 38

Using stigmergy to incorporate the time into artificial neural networks

University of Cambridge Engineering Part IB Information Engineering Elective

The synaptic Basis for Learning and Memory: a Theoretical approach

Active Control of Spike-Timing Dependent Synaptic Plasticity in an Electrosensory System

Zoo400 Exam 1: Mar 25, 1999

International Journal of Advanced Computer Technology (IJACT)

CAS Seminar - Spiking Neurons Network (SNN) Jakob Kemi ( )

Chapter 2: Cellular Mechanisms and Cognition

Basics of Computational Neuroscience

Nervous System Communication. Nervous System Communication. The First Nerve Cells 1/2/11

Biological Psychology

LESSON 3.3 WORKBOOK. Why does applying pressure relieve pain? Workbook. Postsynaptic potentials

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos

Reduction of Overfitting in Diabetes Prediction Using Deep Learning Neural Network

Neurons, Synapses, and Signaling

Chapter 4 Neuronal Physiology

A Connectionist Model based on Physiological Properties of the Neuron

STRUCTURAL ELEMENTS OF THE NERVOUS SYSTEM

Anatomy Review. Graphics are used with permission of: Pearson Education Inc., publishing as Benjamin Cummings (

Computational cognitive neuroscience: 2. Neuron. Lubica Beňušková Centre for Cognitive Science, FMFI Comenius University in Bratislava

BACKPROPOGATION NEURAL NETWORK FOR PREDICTION OF HEART DISEASE

EE 791 Lecture 2 Jan 19, 2015

QUIZ YOURSELF COLOSSAL NEURON ACTIVITY

What is Anatomy and Physiology?

LESSON 3.3 WORKBOOK. Why does applying pressure relieve pain?

A Stochastic and Competitive Hebbian Learning Mechanism through Spike-Timing Dependent Plasticity to Lift Hard Weight Constraints on Hebbian Synapses

Biomimetic Cortical Nanocircuits: The BioRC Project. Alice C. Parker NSF Emerging Models of Technology Meeting July 24, 2008

Questions Addressed Through Study of Behavioral Mechanisms (Proximate Causes)

Nervous System. 2. Receives information from the environment from CNS to organs and glands. 1. Relays messages, processes info, analyzes data

9/28/2016. Neuron. Multipolar Neuron. Astrocytes Exchange Materials With Neurons. Glia or Glial Cells ( supporting cells of the nervous system)

Chapter 11 Introduction to the Nervous System and Nervous Tissue Chapter Outline

Lesson 6 Learning II Anders Lyhne Christensen, D6.05, INTRODUCTION TO AUTONOMOUS MOBILE ROBOTS

All questions below pertain to mandatory material: all slides, and mandatory homework (if any).

You submitted this quiz on Sun 19 May :32 PM IST (UTC +0530). You got a score of out of

Imperfect Synapses in Artificial Spiking Neural Networks

Neural Communication. Central Nervous System Peripheral Nervous System. Communication in the Nervous System. 4 Common Components of a Neuron

You submitted this quiz on Sun 7 Apr :08 PM IST (UTC +0530). You got a score of out of Your Answer Score Explanation

2Lesson. Outline 3.2. Lesson Plan. The OVERVIEW. Lesson 3.2: How do our neurons communicate with each other? LESSON. Unit1.2

Lesson 14. The Nervous System. Introduction to Life Processes - SCI 102 1

Chapter 1: Introduction

Information-theoretic stimulus design for neurophysiology & psychophysics

Neurons, Synapses, and Signaling

MASTER S THESIS. Ali El Adi. Deep neural networks to forecast cardiac and respiratory deterioration of intensive care patients

Electrical Properties of Neurons. Steven McLoon Department of Neuroscience University of Minnesota

Axon initial segment position changes CA1 pyramidal neuron excitability

Teacher Key. Big Idea Different types of neurons compose the nervous tissue that forms the communication system within the body.

APPLICATION OF SPARSE DISTRIBUTED MEMORY TO THE INVERTED PENDULUM PROBLEM

Nervous System. Master controlling and communicating system of the body. Secrete chemicals called neurotransmitters

AU B. Sc.(Hon's) (Fifth Semester) Esamination, Introduction to Artificial Neural Networks-IV. Paper : -PCSC-504

Chapter 7 Nerve Cells and Electrical Signaling

Why did the network make this prediction?

LOGISTIC REGRESSION VERSUS NEURAL

THE HISTORY OF NEUROSCIENCE

Use of artificial neural networks to estimate production variables of broilers breeders in the production phase

Chapter 6 subtitles postsynaptic integration

ARTIFICIAL NEURAL NETWORKS TO DETECT RISK OF TYPE 2 DIABETES

1.4 - Linear Regression and MS Excel

CS 182 Midterm review

Chapter 7. The Nervous System

Predicting Breast Cancer Recurrence Using Machine Learning Techniques

Hebbian Learning in a Random Network Captures Selectivity Properties of Prefrontal Cortex

Neurons, Synapses, and Signaling

Network Models of Frequency Modulated Sweep Detection

Possibilities of Human Memory Simulation for Brain Based Learning Applications

Transcription:

CS 453X: Class 18 Jacob Whitehill

More on k-means

Exercise: Empty clusters (1) Assume that a set of distinct data points { x (i) } are initially assigned so that none of the k clusters is empty. How can it happen that, during execution of the k-means algorithm, a cluster becomes empty?

Exercise: Empty clusters (1) Example: Blue cluster: { 1, 3 } Red cluster: { 2, 4 } 1 4 2 3

Exercise: Empty clusters (1) Example: Blue mean = Red mean 1 4 μ, μ 2 3

Exercise: Empty clusters (1) Example: Blue cluster: { 1, 2, 3, 4 } Red cluster: { } 1 This was arbitrary. It could also be the reverse. 4 μ, μ 2 3

Exercise: Empty clusters (2) Suppose k-means on a set of n distinct data points { x (i) }, where k n, returns at least one cluster that is empty. Can that clustering be globally optimal? kx X x (i) c (j) 2 j=1 i:a(x (i) )=j

Exercise: Empty clusters (2) Suppose k-means on a set of n distinct data points { x (i) }, where k n, returns at least one cluster that is empty. Can that clustering be globally optimal? No. kx j=1 X i:a(x (i) )=j x (i) c (j) 2

Exercise: Empty clusters (2) Reason: Since there are n k data points, and since there were fewer than k clusters returned, then we know there is at least one cluster c with more than 1 data point. Example: n=7, k=3, but one cluster empty.

Exercise: Empty clusters (2) Reason: Consider the data assigned to cluster c. Since they are distinct, then the distance from each point to the cluster mean must be positive. μ

Exercise: Empty clusters (2) Reason: We can reduce the overall cost by creating another cluster (so we have k total) by arbitrarily assigning one of the data in c to the new cluster.

Neural networks

Neural networks Since ~2008, the ML model that has made the largest impact on the field is the (resurgence of) the neural network. Artificial neural networks (NN) have existed since the 1940s. Over the years, their prominence has waxed and waned, as scientists have alternately made new breakthroughs or run into new roadblocks. The field of deep learning based on much computationally deep neural networks than was previously possible has emerged since around 2008.

Biological neural networks In biological systems, neural networks consist of a network of connected neurons, i.e., brain cells. Neurons consist of several components: Soma (cell body) Dendrites (receptors from other neurons) Axons (transmitters to other neurons) https://en.wikipedia.org/wiki/neuron#/media/file:blausen_0657_multipolarneuron.png

Biological neural networks In biological systems, neural networks consist of a network of connected neurons, i.e., brain cells. Neurons consist of several components: Soma (cell body) Dendrites (receptors from other neurons) Axons (transmitters to other neurons) https://en.wikipedia.org/wiki/neuron#/media/file:blausen_0657_multipolarneuron.png

Biological neural networks In biological systems, neural networks consist of a network of connected neurons, i.e., brain cells. Neurons consist of several components: Soma (cell body) Dendrites (receptors from other neurons) Axons (transmitters to other neurons) https://en.wikipedia.org/wiki/neuron#/media/file:blausen_0657_multipolarneuron.png

Biological neural networks Neurons can transmit electrical potentials to other neurons via chemical neurotransmitters. When the voltage change within one neuron exceeds some threshold, an all-or-none electrical potential is transmitted along the axon to neighboring neurons.

Artificial neural networks We can model this process using an artificial neural network: Incoming signals from other neurons Outgoing signals to other neurons Cell body Dendrites Cell body Axons Cell body Cell body

Artificial neural networks We can model this process using an artificial neural network: Incoming signals from other neurons Outgoing signals to other neurons Cell body Dendrites i1 i2 i3 w1 w2 w3 Cell body Axons Cell body If the sum of the incoming potentials i1,, im, weighted by the synaptic weights w1,, wm, exceeds a threshold, then a new outgoing potential o is transmitted along the axons. Cell body

Artificial neural networks We can model this process using an artificial neural network: Incoming signals from other neurons Dendrites i1 i2 i3 w1 w2 w3 Outgoing signals to other neurons o Axons Cell o body If the sum of the incoming potentials i1,, im, weighted by the synaptic weights w1,, wm, exceeds a threshold, then a new outgoing potential o is transmitted along the axons. o Cell body Cell body Cell body

Artificial neural networks We can model this process using an artificial neural network: Incoming signals from other neurons Dendrites Mathematically: 0 1 mx o = @ i j w j A j=1 where σ is (usually) some non-linear activation function. i1 i2 i3 w1 w2 w3 Outgoing signals to other neurons o Axons Cell o body o Cell body Cell body Cell body

Activation functions One of the key ingredients of artificial neural networks is the choice of activation function σ. In the original Perceptron neural network, (Rosenblatt 1957), σ was the Heaviside step function: (z) z = mx i j w j which directly models the all-or-none firing rule of biological neural networks. j=1

Activation functions Because the Heaviside step function is non-differentiable at 0, it was largely replaced by either a logistic sigmoid: (z) = 1 1 + exp z (z) z = mx i j w j j=1

Activation functions or sometimes with hyperbolic tangent: tanh(z) z = mx i j w j j=1

Activation functions More modern neural networks often use a rectified linear activation function, known as ReLU: ReLU(z) = max{0,z} ReLU(z) z = mx i j w j j=1

NNs as a mathematical function Neural networks can compute mathematical functions if we designate a subset of neurons as the input and a subset of neurons as the output. One common network design is a feed-forward (FF) network, consisting of multiple layers of neurons, each of which feeds to the next layer. Here is a simple example, which is equivalent to linear regression: x1 x2 w1 w2 wm ŷ xm Input layer Output layer

NNs as a mathematical function The input layer x directly transmits the values x1,, xm to the next layer. The output layer ŷ computes the sum of the inputs multiplied by the synaptic weights w. In this network, no activation function was used to transform the weighted sum of incoming potentials to ŷ. x1 x2 w1 w2 wm ŷ xm Input layer Output layer

NNs as a mathematical function This network computes the function: ŷ = f(x) = mx x j w j j=1 x1 x2 w1 w2 wm ŷ xm Input layer Output layer

Example: feed-forward NN More commonly, there is at least one layer between the input and output layers; it is known as a hidden layer. The NN processes the input according to the direction of the arrows Input layer Hidden layer Output layer

Example: feed-forward NN First, we just plug in the input x into the first layer: x1 x2 x3 Input layer Hidden layer Output layer

Example: feed-forward NN Next, we compute the sum of incoming potentials to the neuron in the hidden layer: x1 w1 z = mx x j w j x2 w2 w3 j=1 z x3 Input layer Hidden layer Output layer

Example: feed-forward NN We then pass the z to the activation function σ (which could be the logistic sigmoid, tanh, ReLU, etc.) to get h: x1 x2 x3 w1 w2 w3 h = h (z) Input layer Hidden layer Output layer

Example: feed-forward NN Continuing on, we transmit h to the output layer x1 x2 w1 w2 w3 h = h (z) w4 x3 Input layer Hidden layer Output layer

Example: feed-forward NN to obtain the output ŷ: x1 x2 w1 w2 w3 h w4 ŷ = hw 4 ŷ x3 Input layer Hidden layer Output layer

Example: feed-forward NN In aggregate, our NN below computes the function: 0 ŷ = f(x) =w 4 @ 1 3X x j w j A j=1 x1 x2 w1 w2 w3 h w4 ŷ = hw 4 ŷ x3 Input layer Hidden layer Output layer

Exercise: feed-forward NN What will ŷ be for x = [1 0-2] T, w1=1, w2=2, w3=-1.5, and w4=-1? Assume ReLU is the activation function. 0 ŷ = f(x) =w 4 @ 1 3X x j w j A j=1 x1 x2 w1 w2 w3 h w4 ŷ = hw 4 ŷ x3 Input layer Hidden layer Output layer

Exercise: feed-forward NN What will ŷ be for x = [1 0-2] T, w1=1, w2=2, w3=-1.5, and w4=-1? Assume ReLU is the activation function. ŷ = f(x) =w 4 0 @ x1 x2 x3 1 3X x j w j A j=1 w1 w2 w3 h x1w1 = 1, x2w2 = 0, x3w3 = 3 z = 4 h = ReLU(z) = 4 ŷ = hw4 = 4*-1 = -4 w4 ŷ = hw 4 ŷ Input layer Hidden layer Output layer