Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.

Slides:



Advertisements
Similar presentations
6. Radial-basis function (RBF) networks
Advertisements

EE-M /7: IS L7&8 1/24, v3.0 Lectures 7&8: Non-linear Classification and Regression using Layered Perceptrons Dr Martin Brown Room: E1k
G53MLE | Machine Learning | Dr Guoping Qiu
Longin Jan Latecki Temple University
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Data Mining Classification: Alternative Techniques
Support Vector Machines
Pattern Recognition and Machine Learning: Kernel Methods.
Instance Based Learning
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Pattern Recognition and Machine Learning
Radial-Basis Function Networks CS/CMPE 537 – Neural Networks.
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Simple Neural Nets For Pattern Classification
A Review: Architecture
6/10/ Visual Recognition1 Radial Basis Function Networks Computer Science, KAIST.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2010 Shreekanth Mandayam ECE Department Rowan University.
Radial Basis-Function Networks. Back-Propagation Stochastic Back-Propagation Algorithm Step by Step Example Radial Basis-Function Networks Gaussian response.
Radial Basis Functions
S. Mandayam/ ANN/ECE Dept./Rowan University Artificial Neural Networks ECE /ECE Fall 2008 Shreekanth Mandayam ECE Department Rowan University.
Giansalvo EXIN Cirrincione unit #6 Problem: given a mapping: x  d  t  and a TS of N points, find a function h(x) such that: h(x n ) = t n n = 1,
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Radial Basis Function Networks 표현아 Computer Science, KAIST.
An Illustrative Example
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Prediction Networks Prediction –Predict f(t) based on values of f(t – 1), f(t – 2),… –Two NN models: feedforward and recurrent A simple example (section.
Greg GrudicIntro AI1 Introduction to Artificial Intelligence CSCI 3202: The Perceptron Algorithm Greg Grudic.
A Kernel-based Support Vector Machine by Peter Axelberg and Johan Löfhede.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
CS Instance Based Learning1 Instance Based Learning.
Aula 4 Radial Basis Function Networks
Radial Basis Function (RBF) Networks
Radial Basis Function G.Anuradha.
Last lecture summary.
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Last lecture summary.
Chapter 6-2 Radial Basis Function Networks 1. Topics Basis Functions Radial Basis Functions Gaussian Basis Functions Nadaraya Watson Kernel Regression.
Ch. 11: Optimization and Search Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 some slides from Stephen Marsland, some images.
Radial Basis Function Networks
Where We’re At Three learning rules  Hebbian learning regression  LMS (delta rule) regression  Perceptron classification.
Radial Basis Function Networks
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Neural NetworksNN 11 Neural netwoks thanks to: Basics of neural network theory and practice for supervised and unsupervised.
Artificial Neural Networks Shreekanth Mandayam Robi Polikar …… …... … net k   
Chapter 9 Neural Network.
Radial Basis Function Networks:
Artificial Intelligence Techniques Multilayer Perceptrons.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
CS558 Project Local SVM Classification based on triangulation (on the plane) Glenn Fung.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Today’s Lecture Neural networks Training
Machine Learning Supervised Learning Classification and Regression
Neural Networks Winter-Spring 2014
Data Mining, Neural Network and Genetic Programming
Radial Basis Function G.Anuradha.
Hyperparameters, bias-variance tradeoff, validation
Neuro-Computing Lecture 4 Radial Basis Function Network
Neural Network - 2 Mayank Vatsa
Multilayer Perceptron & Backpropagation
Introduction to Radial Basis Function Networks
Prediction Networks Prediction A simple example (section 3.7.3)
Presentation transcript:

Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin Jan Latecki Temple University

Perceptorn

In RBFN

architecture Input layerHidden layerOutput layer x1x1 x2x2 x3x3 xnxn h1h1 h2h2 h3h3 hmhm f(x) W1W1 W2W2 W3W3 WmWm

Three layers Input layer – Source nodes that connect to the network to its environment Hidden layer – Hidden units provide a set of basis function – High dimensionality Output layer – Linear combination of hidden functions architecture

Radial basis function h j (x) = exp( -(x-c j ) 2 / r j 2 ) f(x) =  w j h j (x) j=1 m Wherec j is center of a region, r j is width of the receptive field architecture

7 Function Approximation with Radial Basis Functions RBF Networks approximate functions using (radial) basis functions as the building blocks.

8 Exact Interpolation RBFs have their origins in techniques for performing exact function interpolation [Bishop, 1995]: – Find a function h(x) such that h(x n ) = t n  n=1,... N Radial Basis Function approach (Powel 1987): – Use a set of N basis functions of the form  (||x-x n ||), one for each point,where  (.) is some non-linear function. – Output: h(x) =  n w n  (||x-x n ||)

9 Exact Interpolation Goal (exact interpolation): – Find a function h(x) such that h(x n ) = t n  n=1,... N Radial Basis Function approach (Powel 1987): – Use a set of N basis functions of the form  (||x-x n ||), one for each point,where  (.) is some non-linear function. – Output: h(x) =  n w n  (||x-x n ||) w 1  (||x 1 -x 1 ||) + w 2  (||x 1 -x 2 ||) w N  (||x 1 -x N ||) = t 1 w 1  (||x 2 -x 1 ||) + w 2  (||x 2 -x 2 ||) w N  (||x 2 -x N ||) = t 2  W = T... w 1  (||x N -x 1 ||) + w 2  (||x N -x 2 ||) w N  (||x N -x N ||)= t N

10 Exact Interpolation

11 Exact Interpolation

12 Due to noise that may be present in the data exact interpolation is rarely useful. By introducing a number of modifications, we arrive at RBF networks: Complexity rather than the size of the data is what is important – Number of the basis functions need not be equal to N Centers need not be constrained by the input Each basis function can have its own adjustable width parameter  Bias parameter may be included in the linear sum.

13 Illustrative Example - XOR Problem 22 11

14 Function Approximation via Basis Functions and RBF Networks Using nonlinear functions, we can convert a nonlinearly separable problem into a linearly separable one. From a function approximation perspective, this is equivalent to implementing a complex function (corresponding to the nonlinearly separable decision boundary) using simple functions (corresponding to the linearly separable decision boundary) Implementing this procedure using a network architecture, yields the RBF networks, if the nonlinear mapping functions are radial basis functions. Radial Basis Functions: – Radial: Symmetric around its center – Basis Functions: Also called kernels, a set of functions whose linear combination can generate an arbitrary function in a given function space.

15 RBF Networks

16 RBF Networks

17 Network Parameters What do these parameters represent?  : The radial basis function for the hidden layer. This is a simple nonlinear mapping function (typically Gaussian) that transforms the d- dimensional input patterns to a (typically higher) H- dimensional space. The complex decision boundary will be constructed from linear combinations (weighted sums) of these simple building blocks. u Ji : The weights joining the first to hidden layer. These weights constitute the center points of the radial basis functions. Also called prototypes of data.  : The spread constant(s). These values determine the spread (extend) of each radial basis function. W jk : The weights joining hidden and output layers. These are the weights which are used in obtaining the linear combination of the radial basis functions. They determine the relative amplitudes of the RBFs when they are combined to form the complex function. ||x-u J ||: the Euclidean distance between the input x and the prototype vector u J. Activation of the hidden unit is determined according to this distance through 

18

19 Training RBF Networks Approach 1: Exact RBF Approach 2: Fixed centers selected at random Approach 3: Centers are obtained from clustering Approach 4: Fully supervised training

20 Training RBF Networks Approach 1: Exact RBF Guarantees correct classification of all training data instances. Requires N hidden layer nodes, one for each training instance. No iterative training is involved: w are obtained by solving a set of linear equations Non-smooth, bad generalization

21 Exact Interpolation

22 Exact Interpolation

23 Too Many Receptive Fields? In order to reduce the artificial complexity of the RBF, we need to use fewer number of receptive fields. Approach 2: Fixed centers selected at random. Use M < N data points as the receptive field centers. Fast but may require excessive centers Approach 3: Centers are obtained from unsupervised learning (clustering). Centers no longer has to coincide with data points This is the most commonly used procedure, providing good results.

24 Approach 2 Approach 3 Approach 3.b

25 Determining the Output Weights through learning (LMS)

26 RBFs for Classification

Homework Problem 4.1, p. 117 Problem 4.2, p. 117