MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Mehran University of Engineering and Technology, Jamshoro Department of Electronic Engineering Neural Networks Feedforward Networks By Dr. Mukhtiar Ali.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Machine Learning Neural Networks
Simple Neural Nets For Pattern Classification
Neural Networks Basic concepts ArchitectureOperation.
Radial Basis Functions
Pattern Association A pattern association learns associations between input patterns and output patterns. One of the most appealing characteristics of.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Neural Networks Marco Loog.
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #30 4/15/02 Neural Networks.
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Neural Networks 2nd Edition Simon Haykin
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Artificial Neural Networks
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
Artificial Neural Networks An Overview and Analysis.
Neural Networks Architecture Baktash Babadi IPM, SCS Fall 2004.
2101INT – Principles of Intelligent Systems Lecture 10.
Neural Network Hopfield model Kim, Il Joong. Contents  Neural network: Introduction  Definition & Application  Network architectures  Learning processes.
Artificial Neural Network Unsupervised Learning
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
NEURAL NETWORKS FOR DATA MINING
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Neural Networks and Fuzzy Systems Hopfield Network A feedback neural network has feedback loops from its outputs to its inputs. The presence of such loops.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
Neural Networks 2nd Edition Simon Haykin
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Machine Learning Supervised Learning Classification and Regression
Neural networks.
CS 388: Natural Language Processing: Neural Networks
Learning in Neural Networks
Real Neurons Cell structures Cell body Dendrites Axon
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
CSE P573 Applications of Artificial Intelligence Neural Networks
Simple learning in connectionist networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
XOR problem Input 2 Input 1
CSE 573 Introduction to Artificial Intelligence Neural Networks
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
Capabilities of Threshold Neurons
Lecture Notes for Chapter 4 Artificial Neural Networks
Neural Networks ICS 273A UC Irvine Instructor: Max Welling
Backpropagation.
Simple learning in connectionist networks
David Kauchak CS158 – Spring 2019
Presentation transcript:

MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks

References: “Neural Networks – A Comprehensive Foundation” Simon Haykin, html systems-neural-networks.htm

Neural Networks Learning Cont: Last time we ended with a description of 3 learning paradigms. Today we will begin by introducing some learning algorithms. A learning algorithm will provide the specific implementation of the weight adjustment (how to determine Δw kj.

Neural Networks Error-Correction Learning: This approach involved minimizing the deviation between actual and desired output. So there is a prescribed desired output for each neuron at time step n ( d k (n) ) and an actual output ( y k (n) ).

Neural Networks Thus the deviation between the actual and desired is given by: And we have m such terms if there are m neurons in the network.

Neural Networks So the formulation of our objective function which will attempt to simultaneously minimize all e k ’s is: Where E is the expected value (recall from statistics).

Neural Networks So clearly, by using the expected value we are assuming that our environment is probabilistic in nature but to actually compute the expected value would require knowledge about just what that probability distribution is. Since we do not have that information, we must alter our objective function to accommodate.

Neural Networks We can eliminate the need for knowledge of our distribution by considering only the instantaneous value of the sum of squared errors given by:

Neural Networks So finally, we come to an update relation like this: Where η is a positive constant that defines the rate of learning. It is a subjective value like our mutation rates etc. A very high η will cause your network to learn very quickly but also may cause divergence in the algorithm. A very low η will result in very slow learning.

Neural Networks So it is clear from our objective function that we want to drive each error term to zero. Looking at our update relation, it is clear that by driving our error terms to zero, we will be driving our changes- in-weights to zero. Thus we have a convergence criteria for our learning algorithm.

Neural Networks Point of interest: A plot of our unabridged objective function (J) vs. the weights in our network produces a multidimensional surface that is: - parabolic if we have linear neurons - multimodal if we have non-linear neurons.

Neural Networks Based on what we just learned, what category does error correction learning fall into (which of our 3 paradigms)?

Neural Networks Hebbian Learning Hebbian learning is based on the concept that when 2 neurons in a brain are near enough to stimulate one another and do so often, then the connection between them grows stronger.

Neural Networks We can generalize this to say that if the input through a particular synapse is usually non-zero during time steps in which the neuron fires, then it is a good input and its corresponding weight should be increased. By the same token, if an input through a particular synapse is usually 0 when the neuron fires, then that synaptic weight is correspondingly weakened.

Neural Networks A B C y A ≠ 0 y B = 0 y c ≠ 0 w CA w CB Here, we increase W CA and decrease W CB

Neural Networks Implementation: So we are saying that our update relation for a synaptic weight is a function not only of the input at that synapse, but also of the output of the neuron.

Neural Networks The general form of the update relation is: This is similar to our error-correction update relation but now, since we have no target value we are considering y k in place of e k.

Neural Networks Especially for a long learning process, a synaptic weight has the potential to grow without bound. To combat this, another factor is typically introduced to limit the growth of synaptic weights. The modified update formulation is: Where α is another positive constant that defines the rate of “Forgetting”.

Neural Networks Based on what we just learned, what category does Hebbian learning fall into (which of our 3 paradigms)?

Neural Networks Network Architectures: There are 3 general NN architectures that I will present. Choice of an architecture is intimately linked with choice of a learning algorithm and both are highly dependent on the problem at hand.

Neural Networks 1 – Fully connected Single layer feed forward network Note that there does not have to be the same number of input nodes and output nodes.

Neural Networks 2 – Fully connected Multi layer feed forward network Notice that here we have a “hidden” layer of neurons. (Hidden b/c not seen by input nodes or end effectors).

Neural Networks 3 – recurrent network (with self-feedback loops). Note that by virtue of the feedback, this network is considered to have hidden nodes.

Neural Networks Feedback increases the dynamical nature of the NN. And because the output of a neuron is usually the result of a non-linear function, the result is increased non-linearity for the net.