A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Multi-Layer Perceptron (MLP)
Beyond Linear Separability
A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Introduction to Neural Networks Computing
1 Image Classification MSc Image Processing Assignment March 2003.
Artificial Neural Networks
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Overview over different methods – Supervised Learning
Simple Neural Nets For Pattern Classification
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Artificial Neural Networks
Artificial Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Aula 4 Radial Basis Function Networks
CS 484 – Artificial Intelligence
Radial-Basis Function Networks
Neural Networks. Plan Perceptron  Linear discriminant Associative memories  Hopfield networks  Chaotic networks Multilayer perceptron  Backpropagation.
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
Parallel Artificial Neural Networks Ian Wesley-Smith Frameworks Division Center for Computation and Technology Louisiana State University
Artificial Neural Network Yalong Li Some slides are from _24_2011_ann.pdf.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Lecture No. 29 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Neural Networks Dr. Thompson March 19, Artificial Intelligence Robotics Computer Vision & Speech Recognition Expert Systems Pattern Recognition.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Non-Bayes classifiers. Linear discriminants, neural networks.
11 1 Backpropagation Multilayer Perceptron R – S 1 – S 2 – S 3 Network.
CS621 : Artificial Intelligence
Chapter 2 Single Layer Feedforward Networks
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
EEE502 Pattern Recognition
Previous Lecture Perceptron W  t+1  W  t  t  d(t) - sign (w(t)  x)] x Adaline W  t+1  W  t  t  d(t) - f(w(t)  x)] f’ x Gradient.
Learning: Neural Networks Artificial Intelligence CMSC February 3, 2005.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural networks.
Back Propagation and Representation in PDP Networks
Fall 2004 Backpropagation CS478 - Machine Learning.
第 3 章 神经网络.
CSE 473 Introduction to Artificial Intelligence Neural Networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
CS621: Artificial Intelligence
Synaptic DynamicsII : Supervised Learning
of the Artificial Neural Networks.
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Networks Chapter 5
Neural Network - 2 Mayank Vatsa
Artificial Intelligence Chapter 3 Neural Networks
Backpropagation.
Backpropagation.
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Introduction to Neural Networks
Prof. Pushpak Bhattacharyya, IIT Bombay
CS621: Artificial Intelligence Lecture 18: Feedforward network contd
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -

Solving the differential equation: or in the general form: What is the solution of this type of equation: Try:

THE PERCEPTRON: (Classification) Threshold unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 x1x2y AND

x1x2y AND Linearly seprable

x1x2y OR Linearly separable

Perceptron learning rule: w 1 w 2 w 3 w 4 w 5 Convergence proof: Hertz, Krough, Palmer (HKP) - did you receive the ? Assignment 3a: program in matlab a preceptron with a perceptron learning rule and solve the OR, AND and XOR problems. (Due before Feb 27) Show Demo

Summary – what can perceptrons do and how?

Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error: or *

Linear single layer network: ( approximation, curve fitting) Linear unit: where is the output for input pattern, are the synaptic weights and is the desired output w 1 w 2 w 3 w 4 w 5 Minimize mean square error:

The best solution is obtained when E is minimal. For linear neurons there is an exact solution for this called the pseudo-inverse (see HKP). Looking for a solution by gradient descent: E w -gradient Chain rule

and Since: Error: Therefore: Which types of problems can a linear network solve?

Sigmoidal neurons: Which types of problems can a sigmoidal networks solve? Assignment 3b – Implement a one layer linear and sigmoidal network, fit a 1D a linear, a sigmoid and a quadratic function, for both networks. for example:

Multi layer networks: Can solve non linearly separable classification problems. Can approximate any arbitrary function, given ‘enough’ units in the hidden layer. Hidden layer Output layer Input layer

Note: is not a vector but a matrix

Solving linearly inseparable problems x1x2y XOR Hint: XOR = or and not and

How do we learn a multi-layer network The credit assignment problem ! x1x2y XOR

Gradient descent/ Back Propagation, the solution to the credit assignment problem: Where: From hidden layer to output weights: {

Where: For input to hidden layer: and {

For input to hidden layer: and Assignment 3c: Program a 2 layer network in matlab, solve the XOR problem. Fit the curve: x(x- 1) between 0 and 1, how many hidden units did you need?

Formal neural networks can accomplish many tasks, for example: Perform complex classification Learn arbitrary functions Account for associative memory Some applications: Robotics, Character recognition, Speech recognition, Medical diagnostics. This is not Neuroscience, but is motivated loosely by neuroscience and carries important information for neuroscience as well. For example: Memory, learning and some aspects of development are assumed to be based on synaptic plasticity.

What did we learn today? Is BackProp biologically realistic?