Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.

Slides:



Advertisements
Similar presentations
Multi-Layer Perceptron (MLP)
Advertisements

Beyond Linear Separability
NEURAL NETWORKS Biological analogy
A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Mehran University of Engineering and Technology, Jamshoro Department of Electronic Engineering Neural Networks Feedforward Networks By Dr. Mukhtiar Ali.
Kostas Kontogiannis E&CE
Artificial Neural Networks
Machine Learning Neural Networks
Artificial Neural Networks
Simple Neural Nets For Pattern Classification
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
The back-propagation training algorithm
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Chapter 6: Multilayer Neural Networks
Data Mining with Neural Networks (HK: Chapter 7.5)
CS 4700: Foundations of Artificial Intelligence
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Computer Science and Engineering
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 8: Neural Networks.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Akram Bitar and Larry Manevitz Department of Computer Science
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
EEE502 Pattern Recognition
Neural Networks 2nd Edition Simon Haykin
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Neural networks.
Neural Networks.
Learning with Perceptrons and Neural Networks
Artificial neural networks:
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Prof. Carolina Ruiz Department of Computer Science
Machine Learning Today: Reading: Maria Florina Balcan
of the Artificial Neural Networks.
Neural Network - 2 Mayank Vatsa
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008

© Prof. Rolf Ingold 2 Outline  Biological vs. artificial neural networks  Artificial neuron model  Artificial neural networks  Multi-layer perceptron  Feed-forward activation  Learning approach  Back-propagation method  Optimal learning  Illustration of JavaNNS

© Prof. Rolf Ingold 3 Biological neurons  Artificial neural networks are inspired by biological neurons of the central nervous system  each neuron is connected to many other neurons  information is transmitted via synapses (electro- chemical process)‏  a neuron receives input from its dendrites, and transmit output via the axon to synapses

© Prof. Rolf Ingold 4 Biological vs artificial networks up to 10 8 approx number de synapses up to 10 6 approx number of neurons very fastrelatively slowtransmission time mathematical function chemicalprocessing artificial neural network biological neural network

© Prof. Rolf Ingold 5 Artificial neuron model  A neuron receives input signals x 1,..., x n  These signals are multiplied by synaptic weights w 1,..., w n, which can be positive or negative  The activation of the neuron is transmitted to a non linear function f with threshold w 0  The output signal y = f (a-w 0 ) is then propagated to other neurons

© Prof. Rolf Ingold 6 Characteristics of artificial neural networks  Artificial neural networks may vary in different aspects  the topology of the network, i.e.  the number of neurons, possibly organized in layers or classes  how each neuron (of a given layer/class) is connected to its neighbors  the transfer function used in each neuron  The use and the learning strategy has to be adapted

© Prof. Rolf Ingold 7 Topology of the neural network  The synaptic connections have a major influence on the behavior of the neural network  Two main categories can be considered  feed-forward networks where each neuron is propagating its output signal to neurons that have not yet been used  as special case, the multi-layer perceptron has a sequence of layers such than a neuron from one layer is connected only to neurons of the next layer  dynamic networks where neurons are connected without restrictions, in a cyclic way

© Prof. Rolf Ingold 8 Multi-layer perceptron  The multi-layer perceptron (MLP) has 3 (or more) layers  an input layer with one input neuron per feature  one or several hidden layers having each an arbitrary number of neurons, connected to the previous layer  an output layer with one neuron per class each neuron being connected to the previous layer  Hidden and output layers can be completely or only partly connected  The decision is in favor of the class corresponding to the highest output activation

© Prof. Rolf Ingold 9 Impact of the hidden layer(s)‏  Networks with hidden layers generate arbitrary decision boundaries  however the number of hidden layers has no impact !

© Prof. Rolf Ingold 10 Feed-forward activation  As for the single perceptron, the feature space is augmented with a feature x 0 =1 to take into account the bias w 0.  Each neuron j of a hidden layer computes an activation with  Each neuron k of an output layer computes an activation with

© Prof. Rolf Ingold 11 Transfer function  The transfer function f is supposed to be  monotonic increasing, within the range [-1,+1]  antisymmetric, i.e. f (-net) = - f (net)‏  continuous and derivable (for back-propagation)‏  Typical functions are  step (simple threshold)‏  ramp  sigmoid

© Prof. Rolf Ingold 12 Learning in a multi-layer perceptron  Learning consists of setting the weights w, based on training samples  The method is called back-propagation, because the training error is propagated recursively from the output layer back to the hidden and input layers  The training error on a given pattern is defined as the squared difference between the desired output and the observed output, i.e.  In practice, the desired output is +1 for the correct class and  1 (or sometimes 0 ) for all other classes

© Prof. Rolf Ingold 13 Back-propagation of errors  The weight vectors are changed in the direction of their gradient where  is the learning rate

© Prof. Rolf Ingold 14 Error correction on the output layer  Since the error does not directly depend upon w ji we apply the differential chain rule with and  Thus the update rule becomes

© Prof. Rolf Ingold 15 Error correction on the hidden layer(s)‏  Applying the following chain rule with  Finally the update rule becomes

© Prof. Rolf Ingold 16 Learning algorithm  The learning process starts with randomly initialized weights  The weights are adjusted iteratively by patterns from the training set  the pattern is presented to the network and the feed-forward activation is computed  the output error is computed  the error is used to update the weights reversely, from the output layer to the hidden layers  The process is repeated until a quality criteria is reached

© Prof. Rolf Ingold 17 Risk of overfitting  Minimizing the global error over all training sample tends to produce overfitting  To avoid overfitting, the best strategy is to minimize the global error on a validation set which is independent of the training set

© Prof. Rolf Ingold 18 JavaNNS  JavaNNS is an interactive software framework for experimenting artificial neural networks,  it has been developed at University of Tübingen  it is based on SNNS, an efficient ANN kernel written in C  It supports the following features  multiple topologies (MLP, dynamic networks,...)‏  various transfer functions  various learning strategies  network pruning ...

© Prof. Rolf Ingold 19 Font recognition with JavaNNS  Original neural network with 9 hidden units

© Prof. Rolf Ingold 20 Pruned neural network for font recognition  Neural network obtained after pruning