Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, 281-294 (1989)

Slides:



Advertisements
Similar presentations
NEURAL NETWORKS Backpropagation Algorithm
Advertisements

Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Face Recognition Using Face Unit Radial Basis Function Networks Ben S. Feinstein Harvey Mudd College December 1999.
Instance Based Learning
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Neural Nets Using Backpropagation Chris Marriott Ryan Shirley CJ Baker Thomas Tannahill.
INTRODUCTION TO Machine Learning 2nd Edition
Radial-Basis Function Networks CS/CMPE 537 – Neural Networks.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
6/10/ Visual Recognition1 Radial Basis Function Networks Computer Science, KAIST.
Radial Basis-Function Networks. Back-Propagation Stochastic Back-Propagation Algorithm Step by Step Example Radial Basis-Function Networks Gaussian response.
Radial Basis Functions
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Un Supervised Learning & Self Organizing Maps Learning From Examples
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
Radial Basis Function Networks 표현아 Computer Science, KAIST.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Neural Networks Chapter Feed-Forward Neural Networks.
Artificial Neural Networks
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
CS Instance Based Learning1 Instance Based Learning.
Aula 4 Radial Basis Function Networks
September 28, 2010Neural Networks Lecture 7: Perceptron Modifications 1 Adaline Schematic Adjust weights i1i1i1i1 i2i2i2i2 inininin …  w 0 + w 1 i 1 +
Neural Networks Lecture 17: Self-Organizing Maps
Radial Basis Function (RBF) Networks
Last lecture summary.
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Last lecture summary.
Radial Basis Function Networks
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
CZ5225: Modeling and Simulation in Biology Lecture 5: Clustering Analysis for Microarray Data III Prof. Chen Yu Zong Tel:
Artificial Neural Networks Shreekanth Mandayam Robi Polikar …… …... … net k   
Artificial Neural Networks Dr. Abdul Basit Siddiqui Assistant Professor FURC.
Chapter 9 Neural Network.
RBF TWO-STAGE LEARNING NETWORKS: EXPLOITATION OF SUPERVISED DATA IN THE SELECTION OF HIDDEN UNIT PARAMETERS An application to SAR data classification.
Neural Network Introduction Hung-yi Lee. Review: Supervised Learning Training: Pick the “best” Function f * Training Data Model Testing: Hypothesis Function.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Neural Networks - Lecture 81 Unsupervised competitive learning Particularities of unsupervised learning Data clustering Neural networks for clustering.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. A fast nearest neighbor classifier based on self-organizing incremental neural network (SOINN) Neuron.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
EE459 Neural Networks Backpropagation
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
EEE502 Pattern Recognition
Machine learning optimization Usman Roshan. Machine learning Two components: – Modeling – Optimization Modeling – Generative: we assume a probabilistic.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Today’s Lecture Neural networks Training
Machine Learning Supervised Learning Classification and Regression
Neural Networks Winter-Spring 2014
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
CSC 578 Neural Networks and Deep Learning
Computational Intelligence
Neuro-Computing Lecture 4 Radial Basis Function Network
CSE 573 Introduction to Artificial Intelligence Neural Networks
Computational Intelligence
Computational Intelligence
Artificial Neural Networks
Presentation transcript:

Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)

Network Architecture Responses of neurons are “locally-tuned” or “selective” for some part of the input space. Contains a single hidden layer of these locally-tuned neurons. Hidden layer outputs are fed to a layer of linear neurons, giving network output. For mathematical simplicity, we’ll assume only one neuron in the linear output layer.

Network Architecture (2)

Biological Plausibility Cochlear stereocilia cells in human ear exhibit locally-tuned response to frequency. Cells in visual cortex respond selectively to stimulation that is both local in retinal position and local in angle of orientation. Prof. Wang showed locally-tuned responses to motion of particular speeds and orientations.

Mathematical Definitions A network of M locally-tuned units has overall response function: Here, is a real-valued vector in input space, is the response function of the locally-tuned unit, R is a radially-symmetric function with a single maximum its center and which drops to zero at large radii.

Mathematical Definitions (2) and are the center and width in the input space of the unit, and is the weight or amplitude of the unit. A simple R is the unit normalized Gaussian

Possible Training Methods Fully supervised training to find neuron centers, widths, and amplitude. –Uses error gradient found by varying all parameters (no restrictions on the parameters). –In particular, widths can grow large, thereby losing the local nature of the neurons. –Compared with backpropagation, achieves lower error, but like BP, very slow to train.

Possible Training Methods (2) Combination of supervised and unsupervised learning, a better choice? –Neuron centers and widths are determined through unsupervised learning. –Weights or amplitudes for hidden layer outputs are determined through supervised training.

Unsupervised Learning Determination of neuron centers, how? k-means clustering –Find set a k neuron centers which represent a local minimum of the total squared euclidean distances between the training vectors and the neuron centers. Learning Vector Quantization (LVQ)

Unsupervised Learning (2) Determination of neuron widths, how? P nearest-neighbor heuristics –Vary widths to achieve certain amount of response overlap between each neuron and its P nearest neighbors. Global first nearest-neighbor, P = 1 –Uses global average width between each neuron and its nearest neighbor as net’s uniform width.

Supervised Learning Determination of weights, how? Simple case for 1 linear output –Use Widrow-Hoff learning rule. For a layer of linear outputs? –Simply use Gradient Descent learning rule. Reduced to a linear optimization problem.

Advantages Over Backprop Training via a combination of linear supervised and linear self-organizing techniques is much faster than backprop. For a given input, only a small fraction of neurons (those with nearby centers) will give ~ non-zero responses. Hence we don’t need to fire all neurons to get overall output. This improves performance.

Advantages Over Backprop (2) Based on well-developed mathematical theory (kernel theory) yielding statistical robustness. Computational simplicity since only one layer is involved in supervised training. Provides guaranteed, globally optimal solution via simple linear optimization.

Project Proposal Currently debugging C++ RBF network with n dimensional input and 1 linear output neuron. –Uses k-means clustering, global first nearest neighbor heuristic, and gradient descent. –Experimentation with different training algs. Try to reproduce results for RBF neural nets performing face-recognition.