NNs Adaline 1 Neural Networks - Adaline L. Manevitz.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Multi-Layer Perceptron (MLP)
Beyond Linear Separability
Slides from: Doug Gray, David Poole
Artificial Neural Networks
Classification Neural Networks 1
Perceptron.
Overview over different methods – Supervised Learning
Artificial Neural Networks
Lecture 14 – Neural Networks
Simple Neural Nets For Pattern Classification
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Learning via Neural Networks L. Manevitz All rights reserved.
Back-Propagation Algorithm
Neural Network Computing Lecture no.1. All rights reserved L. Manevitz Lecture 12 McCullogh-Pitts Neuron The activation of a McCullogh-Pitts Neuron is.
Artificial Neural Networks
Before we start ADALINE
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Artificial Neural Networks
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Neural Networks Lecture 8: Two simple learning algorithms
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Artificial Neural Networks
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
1 Artificial Neural Networks Sanun Srisuk EECP0720 Expert Systems – Artificial Neural Networks.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent. x0x0 + -
Artificial Intelligence Chapter 3 Neural Networks Artificial Intelligence Chapter 3 Neural Networks Biointelligence Lab School of Computer Sci. & Eng.
Computational Modeling of Cognitive Activity Prof. Larry M. Manevitz Course Slides: 2012.
Non-Bayes classifiers. Linear discriminants, neural networks.
L. Manevitz U. Haifa 1 Neural Networks: Capabilities and Examples L. Manevitz Computer Science Department HIACS Research Center University of Haifa.
ADALINE (ADAptive LInear NEuron) Network and
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Neural NetworksNN 21 Architecture We consider the architecture: feed- forward NN with one layer It is sufficient to study single layer perceptrons with.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Today’s Lecture Neural networks Training
One-layer neural networks Approximation problems
第 3 章 神经网络.
Ranga Rodrigo February 8, 2014
CSE 473 Introduction to Artificial Intelligence Neural Networks
Derivation of a Learning Rule for Perceptrons
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Wed June 12 Goals of today’s lecture. Learning Mechanisms
CSE P573 Applications of Artificial Intelligence Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Data Mining with Neural Networks (HK: Chapter 7.5)
Classification Neural Networks 1
Chapter 3. Artificial Neural Networks - Introduction -
Synaptic DynamicsII : Supervised Learning
Computational Modeling of Cognitive Activity
Artificial Intelligence Chapter 3 Neural Networks
Learning via Neural Networks
CSE 573 Introduction to Artificial Intelligence Neural Networks
Neural Network - 2 Mayank Vatsa
Lecture Notes for Chapter 4 Artificial Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Chapter - 3 Single Layer Percetron
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

NNs Adaline 1 Neural Networks - Adaline L. Manevitz

NNs Adaline 2 Plan of Lecture Perceptron: Connected and Convex examples Adaline Square Error Gradient Calculate for and, xor Discuss limitations LMS algorithm: derivation.

NNs Adaline 3 What is best weights? Most classified correctly? Least Square Error Least Square Error Before Cut-Off! To Minimize  d –  w x)**2 (first sum over examples; second over dimension)

NNs Adaline 4 Least Square Minimization Find gradient of error over all examples. Either calculate the minimum or move opposite to gradient. Widrow-Hoff(LMS): Use instantaneous example as approximation to gradient. –Advantages: No memory; on-line; serves similar function as noise to avoid local problems. –Adjust by w (new) = w (old) +  x for each x. –Here  desired output –  wx )

NNs Adaline 5 LMS Derivation Errsq =  d(k) – W x(k)) ** 2 Grad(errsq) = 2  d(k) – W x(k)) (-x(k)) W (new) = W(old) -  Grad(errsq) To ease calculations, use Err(k) in place of Errsq W(new) = W(old) + 2  Err(k) x(k)) Continue with next choice of k

NNs Adaline 6 Applications Adaline has better convergence properties than Perceptron Useful in noise correction Adaline in every modem.

NNs Adaline 7 LMS (Least Mean Square Alg.) 1. Apply input to Adaline input 2. Find the square error of current input –Errsq(k) = (d(k) - W x(k))**2 3. Approximate Grad(ErrorSquare) by –differentiating Errsq –approximating average Errsq by Errsq(k) –obtain -2Errsq(k)x(k) 4.Update W: W(new) = W(old) + 2  Errsq(k)X(k) 5. Repeat steps 1 to 4.

NNs Adaline 8 Comparison with Perceptron Both use updating rule changing with each input One fixes binary error; the other minimizes continuous error Adaline always converges; see what happens with XOR Both can REPRESENT Linearly separable functions

NNs Adaline 9 Convergence Phenomenom LMS converges depending on choice of  How to choose it?

NNs Adaline 10 Limitations Linearly Separable How can we get around it? –Use network of neurons? –Use a transformation of data so that it is linearly separable

NNs Adaline 11 Multi-level Neural Networks Representability –Arbitrarily complicated decisions –Continuous Approximation: Arbitrary Continuous Functions (and more) (Cybenko Theorem) Learnability –Change Mc-P neurons to Sigmoid etc. –Derive backprop using chain rule. (Like LMS TheoremSample Feed forward Network (No loops)

NNs Adaline 12 Replacement of Threshold Neurons with Sigmoid or Differentiable Neurons Threshold Sigmoid

NNs Adaline 13 Prediction Input/Output NN delay Compare

NNs Adaline 14 Sample Feed forward Network (No loops) Weights Input Output Wji Vik F(  wji xj