Each neuron has a threshold value Each neuron has weighted inputs from other neurons The input signals form a weighted sum If the activation level exceeds.

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

NEURAL NETWORKS Backpropagation Algorithm
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Artificial Neural Networks
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Reading for Next Week Textbook, Section 9, pp A User’s Guide to Support Vector Machines (linked from course website)
Machine Learning Neural Networks
Artificial Neural Networks
Lecture 14 – Neural Networks
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Introduction to Neural Networks Simon Durrant Quantitative Methods December 15th.
Chapter 6: Multilayer Neural Networks
Artificial Neural Networks
Foundations of Learning and Adaptive Systems ICS320
CS 484 – Artificial Intelligence
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial Neural Networks - Introduction -. Overview 1.Biological inspiration 2.Artificial neurons and neural networks 3.Application.
Artificial neural networks:
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Using Neural Networks in Database Mining Tino Jimenez CS157B MW 9-10:15 February 19, 2009.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
© Negnevitsky, Pearson Education, Will neural network work for my problem? Will neural network work for my problem? Character recognition neural.
Multi-Layer Perceptrons Michael J. Watts
Artificial Neural Networks An Overview and Analysis.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
Waqas Haider Khan Bangyal. Multi-Layer Perceptron (MLP)
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University EE459 Neural Networks The Structure.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
NEURAL NETWORKS FOR DATA MINING
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Neural Networks and Machine Learning Applications CSC 563 Prof. Mohamed Batouche Computer Science Department CCIS – King Saud University Riyadh, Saudi.
Akram Bitar and Larry Manevitz Department of Computer Science
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
SUPERVISED LEARNING NETWORK
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
EEE502 Pattern Recognition
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Essential components of the implementation are:  Formation of the network and weight initialization routine  Pixel analysis of images for symbol detection.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks Lecture 4 out of 4. Practical Considerations Input Architecture Output.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Neural Networks An Introduction
Neural Networks Dr. Peter Phillips.
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Artificial Intelligence Methods
network of simple neuron-like computing elements
Neural Network - 2 Mayank Vatsa
McCulloch–Pitts Neuronal Model :
Computer Vision Lecture 19: Object Recognition III
Artificial Neural Networks
A task of induction to find patterns
Presentation transcript:

Each neuron has a threshold value Each neuron has weighted inputs from other neurons The input signals form a weighted sum If the activation level exceeds the threshold, the neuron “fires” The McCulloch-Pitts “Unit” The single neuron (Perceptron)

What is the output of this perceptron? W0 = 1.1 W1 = 0.6 W2 = 0.7 W3 = 0.8

MultiLayer Networks

Feed Forward Network

Types of Layers The input layer. –Introduces input values into the network. –No activation function or other processing. The hidden layer(s). –Perform classification of features The output layer. –Functionally just like the hidden layers –Outputs are passed on to the world outside the neural network.

Feed Forward Network Given the following weights: W0,3 = -1.5 W1,3 = - 1 W2,3 = -1 W0,4 = 1.0 W1,4 = 0.5 W2,4 = 1.5 W0,5 = 1.5 W3,5 = 1 W4.5 = 1 What is the output given the following inputs I1I2HN3HN4O

Feed Forward Network Given the following weights: W0,3 = -1.5 W1,3 = - 1 W2,3 = -1 W0,4 = 1.0 W1,4 = 0.5 W2,4 = 1.5 W0,5 = 1.5 W3,5 = 1 W4.5 = 1 What is the output given the following inputs I1I2HN3HN4O

Feed Forward Network Given the following weights: W0,3 = -1.5 W1,3 = - 1 W2,3 = -1 W0,4 = 1.0 W1,4 = 0.5 W2,4 = 1.5 W0,5 = 1.5 W3,5 = 1 W4.5 = 1 What is the output given the following inputs I1I2HN3HN4O

Feed Forward Network Given the following weights: W0,3 = -1.5 W1,3 = - 1 W2,3 = -1 W0,4 = 1.0 W1,4 = 0.5 W2,4 = 1.5 W0,5 = 1.5 W3,5 = 1 W4.5 = 1 What is the output given the following inputs I1I2HN3HN4O

Feed Forward Network Given the following weights: W0,3 = -1.5 W1,3 = - 1 W2,3 = -1 W0,4 = 1.0 W1,4 = 0.5 W2,4 = 1.5 W0,5 = 1.5 W3,5 = 1 W4.5 = 1 What is the output given the following inputs I1I2HN3HN4O

How many layers? It depends on the problem…

How many layers? It depends on the problem… The hidden layer(s). –Perform classification of features –Two hidden layers are sufficient to solve any problem –More features imply more layers may be better

Different Non-Linearly Separable Problems Structure Types of Decision Regions Exclusive-OR Problem Classes with Meshed regions Most General Region Shapes Single-Layer Two-Layer Three-Layer Half Plane Bounded By Hyperplane Convex Open Or Closed Regions Arbitrary (Complexity Limited by No. of Nodes) A AB B A AB B A AB B B A B A B A

Hyperplane partitions An extra layer models a convex hull ( “An area with no dents in it”) –Two layers add convex hulls together –Sufficient to classify anything “sane”. In theory, further layers add nothing In practice, extra layers may be better

Training Algorithms Adjust neural network weights to map inputs to outputs. Use a set of sample patterns where the desired output (given the inputs presented) is known. The purpose is to learn to generalize –Recognize features which are common to good and bad exemplars

Back-Propagation A training procedure which allows multi-layer feedforward Neural Networks to be trained; Can theoretically perform “any” input-output mapping; Can learn to solve linearly inseparable problems.

Applications The properties of neural networks define where they are useful. –Can learn complex mappings from inputs to outputs, based solely on samples –Difficult to analyse: firm predictions about neural network behaviour difficult; Unsuitable for safety-critical applications. –Require limited understanding from trainer, who can be guided by heuristics.

Applications OCR Engine Management Navigation Signature Recognition Sonar Recognition Stock Market Prediction Mortgage Assessment

Neural network for OCR feedforward network trained using Back- propagation

OCR for 8x10 characters NN are able to generalise learning involves generating a partitioning of the input space for single layer network input space must be linearly separable what is the dimension of this input space? how many points in the input space? this network is binary (uses binary values) networks may also be continuous

Engine management The behaviour of a car engine is influenced by a large number of parameters –temperature at various points –fuel/air mixture –lubricant viscosity. Major companies have used neural networks to dynamically tune an engine depending on current settings.

ALVINN Drives 70 mph on a public highway 30x32 pixels as inputs 30 outputs for steering 30x32 weights into one out of four hidden unit 4 hidden units

Signature recognition Each person's signature is different. There are structural similarities which are difficult to quantify. One company has manufactured a machine which recognizes signatures to within a high level of accuracy. –Considers speed in addition to gross shape. –Makes forgery even more difficult.

Sonar target recognition Distinguish mines from rocks on sea-bed The neural network is provided with a large number of parameters which are extracted from the sonar signal. The training set consists of sets of signals from rocks and mines.

Stock market prediction “Technical trading” refers to trading based solely on known statistical parameters; e.g. previous price Neural networks have been used to attempt to predict changes in prices. Difficult to assess success since companies using these techniques are reluctant to disclose information.

Mortgage assessment Assess risk of lending to an individual. Difficult to decide on marginal cases. Neural networks have been trained to make decisions, based upon the opinions of expert underwriters. Neural network produced a 12% reduction in delinquencies compared with human experts.

Neural Network Problems Many Parameters to be set Overfitting long training times...

Parameter setting Number of layers Number of neurons too many neurons, require more training time Learning rate from experience, value should be small ~0.1

Over-fitting With sufficient nodes can classify any training set exactly May have poor generalization ability. Cross-validation with some patterns –Typically 30% of training patterns –Validation set error is checked each epoch –Stop training if validation error goes up

Training time How many epochs of training? –Stop if the error fails to improve (has reached a minimum) –Stop if the rate of improvement drops below a certain level –Stop if the error reaches an acceptable level –Stop when a certain number of epochs have passed

So what would an ANN look like?

Remember what the Tree Looked Like? Outlook SunnyOvercastRain Humidity HighNormal Wind StrongWeak NoYes No [D3,D7,D12,D13] [D8,D9,D11] [D6,D14] [D1,D2] [D4,D5,D10]

Forecast = Overcast Forecast = Sunny Humidity=High Forecast = Rain Wind = Weak W 0 = 1.5 W 0 = 0.5

Forecast = Overcast Forecast = Sunny Humidity=High Forecast = Rain Wind = Weak W 0 = 0 W 0 = 1.5 W 0 = 0.5

Forecast = Overcast Forecast = Sunny Humidity=High Forecast = Rain Wind = Weak W 0 = 0 W 0 = 1.5 W 0 = 0.5