U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d www.usq.edu.au Neural Networks and Self-Organising Maps CSC3417 Semester 1, 2007.

Slides:



Advertisements
Similar presentations
An Introduction to Calculus. Calculus Study of how things change Allows us to model real-life situations very accurately.
Advertisements

Aula 3 Single Layer Percetron
Integrated photonic beamformer employing
The Derivative in Graphing and Application
Calculating Slope m = y2 – y1 x2 – x1.
Artificial Intelligence 12. Two Layer ANNs
Maximal Independent Subsets of Linear Spaces. Whats a linear space? Given a set of points V a set of lines where a line is a k-set of points each pair.
Stress Matrix Visualization Wednesday, 9/4/2002. Stress Vector.
Multi-Layer Perceptron (MLP)
Findings from nurse visit data Mai Stafford, Paola Zaninotto, Cesar de Oliveira, Meena Kumari, Andrew Steptoe Department of Epidemiology & Public Health,
Neural Networks: An Overview. There are many different kinds of Neural Network they include to name a few: The Back Propagating Networks Kohonen (Self-Organising)
NN – cont. Alexandra I. Cristea USI intensive course Adaptive Systems April-May 2003.
Cellular Networks.
Feedback Networks and Associative Memories
Genome 559: Introduction to Statistical and Computational Genomics Elhanan Borenstein Artificial Neural Networks Some slides adapted from Geoffrey Hinton.
1 Machine Learning: Lecture 1 Overview of Machine Learning (Based on Chapter 1 of Mitchell T.., Machine Learning, 1997)
1 Machine Learning: Connectionist Introduction 11.1Foundations of Connectionist Networks 11.2Perceptron Learning 11.3Backpropagation Learning 11.4.
VERILOG: Synthesis - Combinational Logic Combination logic function can be expressed as: logic_output(t) = f(logic_inputs(t)) Rules Avoid technology dependent.
Programming for GCSE Topic 9.1: Logic Circuits T eaching L ondon C omputing William Marsh School of Electronic Engineering and Computer Science Queen Mary.
x1x1 x2x2 (X 1,X 2 ) BY: Grupo CDPYE-UGR This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivs 2.5 License.
How Microsoft Had Made Deep Learning Red-Hot in IT Industry
Light Detectives. Limb of Sun Very low mass star Brown dwarf Jupiter Light Detectives – #2.
Boolean Logic ITI 1121 N. El Kadri. 2 What is a switching network? Switching Network X1X1 XmXm X2X2 Z1Z1 ZmZm Z2Z2 Combinatorial Network: A stateless.
Australian Institute for Teaching and School Leadership Australian Professional Standards for Teachers Unpacking the Standards.
Gradient of a straight line x y 88 66 44 2 44 4 For the graph of y = 2x  4 rise run  = 8  4 = 2 8 rise = 8 4 run = 4 Gradient = y.
 How many neurons are to be used?  How the neurons are to be connected to form a network.  Which learning algorithm to use?  How to train the neural.
Global Constraints Toby Walsh National ICT Australia and University of New South Wales
Computer Graphics … how renderings are done on a computer. Art 321 Dr. J Parker Winter.
Memristor in Learning Neural Networks
2 x0 0 12/13/2014 Know Your Facts!. 2 x1 2 12/13/2014 Know Your Facts!
2 x /18/2014 Know Your Facts!. 11 x /18/2014 Know Your Facts!
Supervised Multiattribute Classification
2 x /10/2015 Know Your Facts!. 8 x /10/2015 Know Your Facts!
Learning from Example Given some data – build a model to make predictions Linear Models (Perceptrons). Support Vector Machines.
What is a neural net? Aziz Kustiyo
Dynamic Software Updates: The State Mapping Problem Rida A. Bazzi Kristis Makris Peyman Nayeri Jun Shen ARIZONA STATE UNIVERSITY.
快速傅里叶变换与多项式乘法优化 赵卓越. 多项式乘法 离散傅里叶变换 (DFT)  定义  逆变换( IDFT )  时间复杂度.
Artificial Neural Networks And XML
1 Lecture 5 PRAM Algorithm: Parallel Prefix Parallel Computing Fall 2008.
5 x4. 10 x2 9 x3 10 x9 10 x4 10 x8 9 x2 9 x4.
Shannon Expansion Given Boolean expression F = w 2 ’ + w 1 ’w 3 ’ + w 1 w 3 Shannon Expansion of F on a variable, say w 2, is to write F as two parts:
0 x x2 0 0 x1 0 0 x3 0 1 x7 7 2 x0 0 9 x0 0.
BINARY/MIXED-INTEGER PROGRAMMING ( A SPECIAL TYPE OF INTEGER PROGRAMMING)
Introduction to Neural Networks 2. Overview  The McCulloch-Pitts neuron  Pattern space  Limitations  Learning.
Power Systems Application of Artificial Neural Networks. (ANN)  Introduction  Brief history.  Structure  How they work  Sample Simulations. (EasyNN)
Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
Presenting: Itai Avron Supervisor: Chen Koren Mid Semester Presentation Spring 2005 Implementation of Artificial Intelligence System on FPGA.
CS5201 Intelligent Systems (2 unit) Semester II Lecturer: Adrian O’Riordan Contact: is office is 312, Kane
Pattern Similarity and Storage Capacity of Hopfield Network Suman K Manandhar Prof. Ramakoti Sadananda Computer Science and Information Management AIT.
Explorations in Neural Networks Tianhui Cai Period 3.
The Perceptron. Perceptron Pattern Classification One of the purposes that neural networks are used for is pattern classification. Once the neural network.
Computer Go : A Go player Rohit Gurjar CS365 Project Presentation, IIT Kanpur Guided By – Prof. Amitabha Mukerjee.
Logan Lebanoff Mentor: Haroon Idrees. Two-layer method  Trying a method that will have two layers of neural networks.
Presented by Matt Brunner
RADIAL BASIS FUNCTION NEURAL NETWORK DESIGN
Data Mining, Neural Network and Genetic Programming
Other Applications of Energy Minimzation
Hire Toyota Innova in Delhi for Outstation Tour
Shadow Depth Map Illumination
كاربردهاي داده كاوي در بانكداري
Supplemental slides for CSE 327 Prof. Jeff Heflin
إعداد المشرفة التربوية نجلاء الجارد
Neural Networks Advantages Criticism
Gene expression profiling diagnosis through DNA molecular computation
Chapter 6: Feedforward mapping networks
Chap 8: Adaptive Networks
Fig. 1. A diagram of artificial neural network consisting of multilayer perceptron. This simple diagram is for a conceptual explanation. When the logistic.
SELF-SERVICE for the CANTINE
Neural Networks…. …. The Undiscovered Country. Jonathan Y
Areas under the receiver operating characteristic (ROC) curves for both the training and testing data sets based on a number of hidden-layer perceptrons.
Presentation transcript:

U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Neural Networks and Self-Organising Maps CSC3417 Semester 1, 2007

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Perceptron (one node neural network) Y = f(  w i x i )Y = f(  w i x i ) + w0w0 w1w1 w2w2 x0x0 x1x1 x2x2 y

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d R Code for a Perceptron step <- function(sigma) {step <- function(sigma) { if (sigma<0) { if (sigma<0) { return (0); return (0); } else { return (1); } } else { return (1); } } neural <- function(weights,inputs,output,eta) {neural <- function(weights,inputs,output,eta) { sigma <- t(inputs) %*% weights; sigma <- t(inputs) %*% weights; yhat <- step(sigma); yhat <- step(sigma); diff <- output-yhat; diff <- output-yhat; for (k in 1:length(weights)) { for (k in 1:length(weights)) { weights[[k]] <- weights[[k]] + eta*inputs[[k]]*diff; weights[[k]] <- weights[[k]] + eta*inputs[[k]]*diff; } } return (weights); return (weights); } Start with random weights and use this function iteratively.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Try it: > w w <- runif(3); > w [1] > w w <- neural(w,c(1,0,1),0,0.1) > w> w [1] [1] > w w <- neural(w,c(1,1,0),1,0.1) > w> w [1] [1] > w w <- neural(w,c(1,1,1),0,0.1) > w> w [1] [1] ……………. > w w <- neural(w,c(1,1,1),0,0.1) > w> w [1] [1] > w w <- neural(w,c(1,1,0),1,0.1) > w> w [1] [1]

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Scilab version function [stp] = step (sigma) if sigma<0 if sigma<0 stp = 0; stp = 0; else else stp = 1; stp = 1; end endendfunction function [w] = neural(weights,inputs,output,eta) sigma = weights * inputs'; sigma = weights * inputs'; yhat = step(sigma); yhat = step(sigma); diff = output-yhat; diff = output-yhat; w = weights; w = weights; for k=1:length(weights) for k=1:length(weights) w(k) = weights(k) + eta*inputs(k)*diff; w(k) = weights(k) + eta*inputs(k)*diff; end endendfunction

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Try it: -->w = rand(1,3) w = w = >w = neural(w,[1,0,1],0,0.1) w = w = >w = neural(w,[1,1,0],1,0.1) w = w = >w = neural(w,[1,1,1],0,0.1) w = w = ………..

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Aspects of Neural Network Use: ArchitectureArchitecture –Input : hidden : output Activation ruleActivation rule –Weights, add, sigmoidal function Learning ruleLearning rule –Correct weights, back-propagation Avoiding over-learningAvoiding over-learning –Stop learning sufficiently early.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Multilayer networks Learning by back-propagation.Learning by back-propagation.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Back Propagation Our objective is to minimise:Our objective is to minimise: Here x is the inputs and w the weights; y are output values, and y hat our estimates.Here x is the inputs and w the weights; y are output values, and y hat our estimates. Back-propagation is a way to compute the gradient of the derivative of this objective function.Back-propagation is a way to compute the gradient of the derivative of this objective function.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d How to avoid overfitting Stop learning earlyStop learning early Reduce the learning speedReduce the learning speed Bayesian approach:Bayesian approach: –“The overfitting problem can be solve by using a Bayesian approach to control model complexity.” MacKay, Information theory, Inference and Learning Algorithms, CUP, 2003.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Advantages of Bayesian Approach (MacKay 2003) 1.No ‘test set’ required – hence more data efficient. 2.Overfitting parameters just appear as ordinary parameters. 3.Bayesian objective function is smooth, hence easier to optimize. 4.Gradient of objective function can be evaluated. McKay’s book is online: ac.uk/mackay/Book.html

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Self-organising Maps Each node has a vector of parameters of the same dimension as the inputs.Each node has a vector of parameters of the same dimension as the inputs. Training moves nodes in the neighbourhood of the closest node to an input closer to each other.Training moves nodes in the neighbourhood of the closest node to an input closer to each other. Many choices of neighbourhood and adjustment are possible.Many choices of neighbourhood and adjustment are possible.

T h e U n i v e r s i t y o f S o u t h e r n Q u e e n s l a n dU n i v e r s i t y o f S o u t h e r n Q u e e n s l a n d Neighbourhood and adjustment where and