Today’s Topics 11/10/15CS 540 - Fall 2015 (Shavlik©), Lecture 21, Week 101 More on DEEP ANNs –Convolution –Max Pooling –Drop Out Final ANN Wrapup FYI:

Slides:



Advertisements
Similar presentations
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
Advertisements

Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Deep Learning and Neural Nets Spring 2015
For Wednesday Read chapter 19, sections 1-3 No homework.
Artificial Neural Networks
Machine Learning Neural Networks
Lecture 14 – Neural Networks
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Artificial Neural Networks ML Paul Scheible.
Neural Networks Marco Loog.
Artificial Neural Networks
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
Face Detection using the Viola-Jones Method
Artificial Neural Networks
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Today’s Topics HW0 due 11:55pm tonight and no later than next Tuesday HW1 out on class home page; discussion page in MoodleHW1discussion page Please do.
Today’s Topics Read –For exam: Chapter 13 of textbook –Not on exam: Sections & Genetic Algorithms (GAs) –Mutation –Crossover –Fitness-proportional.
Today’s Topics HW1 Due 11:55pm Today (no later than next Tuesday) HW2 Out, Due in Two Weeks Next Week We’ll Discuss the Make-Up Midterm Be Sure to Check.
Today’s Topics Playing Deterministic (no Dice, etc) Games –Mini-max –  -  pruning –ML and games? 1997: Computer Chess Player (IBM’s Deep Blue) Beat Human.
Today’s Topics 11/10/15CS Fall 2015 (Shavlik©), Lecture 22, Week 101 Support Vector Machines (SVMs) Three Key Ideas –Max Margins –Allowing Misclassified.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Today’s Topics Midterm class mean: 83.5 HW3 Due Thursday and HW4 Out Thursday Turn in Your BN Nannon Player (in Separate, ‘Dummy’ Assignment) until a Week.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Today’s Topics Read: Chapters 7, 8, and 9 on Logical Representation and Reasoning HW3 due at 11:55pm THURS (ditto for your Nannon Tourney Entry) Recipe.
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Today’s Topics 11/17/15CS Fall 2015 (Shavlik©), Lecture 23, Week 111 Review of the linear SVM with Slack Variables Kernels (for non-linear models)
Intro. ANN & Fuzzy Systems Lecture 13. MLP (V): Speed Up Learning.
Today’s Topics Artificial Neural Networks (ANNs) Perceptrons (1950s) Hidden Units and Backpropagation (1980s) Deep Neural Networks (2010s) ??? (2040s [note.
Neural Networks William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
CSC2535: Computation in Neural Networks Lecture 8: Hopfield nets Geoffrey Hinton.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Intro. ANN & Fuzzy Systems Lecture 11. MLP (III): Back-Propagation.
Machine Learning Supervised Learning Classification and Regression
Presented by Yuting Liu
Deep Learning Amin Sobhani.
cs638/838 - Spring 2017 (Shavlik©), Week 10
Data Mining, Neural Network and Genetic Programming
Computer Science and Engineering, Seoul National University
cs540 - Fall 2015 (Shavlik©), Lecture 25, Week 14
Classification with Perceptrons Reading:
CS Fall 2016 (Shavlik©), Lecture 12, Week 6
Supervised Training of Deep Networks
CS 188: Artificial Intelligence
Convolutional Networks
Deep Belief Networks Psychology 209 February 22, 2013.
Prof. Carolina Ruiz Department of Computer Science
cs638/838 - Spring 2017 (Shavlik©), Week 7
CS540 - Fall 2016 (Shavlik©), Lecture 17, Week 10
Machine Learning Today: Reading: Maria Florina Balcan
Lecture 11. MLP (III): Back-Propagation
CS 4501: Introduction to Computer Vision Training Neural Networks II
cs540 - Fall 2016 (Shavlik©), Lecture 20, Week 11
cs540 - Fall 2016 (Shavlik©), Lecture 18, Week 10
Neural Networks Geoff Hulten.
Lecture Notes for Chapter 4 Artificial Neural Networks
CS Fall 2016 (Shavlik©), Lecture 12, Week 6
cs638/838 - Spring 2017 (Shavlik©), Week 7
Artificial Neural Networks
Artificial Intelligence 10. Neural Networks
Backpropagation David Kauchak CS159 – Fall 2019.
COSC 4335: Part2: Other Classification Techniques
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Word representations David Kauchak CS158 – Fall 2016.
Neural Machine Translation using CNN
CSC 578 Neural Networks and Deep Learning
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Today’s Topics 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 101 More on DEEP ANNs –Convolution –Max Pooling –Drop Out Final ANN Wrapup FYI: Some Resources machine_9.htmlhttp://googleresearch.blogspot.com/2015/11/tensorflow-googles-latest- machine_9.html deep-learning-modules-for-torch/ deep-learning-modules-for-torch

Back to Deep ANNs - Convolution & Max Pooling 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 2 C = Convolution, MP = Max Pooling (next) (ie, a CHANGE OF REP)

Look for 8’s in all the Right Places 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 3 Imagine we have a great 8 detector expressed as an 8x8 array of 0-1’s (see upper left) We want to find all the 8’s in an 1024x1024 image of 0-1’s Q:What might we do? A: ‘Slide’ the detector across the image and count # of matching bits in the detector and the ‘overlaid’ image 888 If count greater than some threshold, say an ‘8’ is there etc

Look for 8’s in all the Right Places 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 4 Q:What about 8’s in the image larger than 8x8 bits? A: Use ‘detectors’ of, say, 16x16, 32x32, 64x64, etc PS:Could also ‘slide’ slightly rotated 8’s of various sizes (too much rotation and becomes infinity symbol!) 8 etc 88

Back to Deep ANNs - Convolution (cont.) –but each ‘template’ is a HU and the wgts are learned –some HUs are coupled –each group of HUs learns what to ‘look for’ –we do hard-code the ‘size’ of the ‘template’ 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 5 The ‘sliding window’ is the basic idea of convolution Input units rep’ing the image HU2 HU1 Our code would employ weight sharing, ie the corresponding weights in each HU (eg, the two thicker lines) would always have the same value Note: HU77, say, would connect to the same INPUTS as HU1 but would have different wgts, ie would be a different ‘feature dectors’

BACKUP: A Possibly Helpful Slide on Convolution from the Web 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 6

Back to Deep ANNs - Max Pooling Researchers have empirically found it helpful to ‘clean up’ the convolution scores by –Creating the next layer of HUs where each HU holds the MAX score in an N  N array, for various values of N and across various locations –This is called MAX POOLING (example on next slide) –Advanced note (not on final): I’m not sure if people (a) use the differentiable ‘soft max’ ( and BP through all nodes or (b) only BP through the max node; I’d guess (b) 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 7

Back to Deep ANNs - Max Pooling Example (connections not shown) 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week Hidden Layer i Possible Nodes in Hidden Layer i + 1 4x4 max x2 max, non overlapping x2 max, overlapping (contains non- overlapping, so no need for both) 9

Back to Deep ANNs - Drop Out (from Hinton’s Group) Each time one example is processed (forward + back prop) during TRAINING Randomly turn off (‘ie, drop out’) a fraction (say,  = ½) of the input and hidden units During TESTING scale all weights by (1 -  ), since that is the fraction of time each unit was present during training (ie, so on average, weighted sums are the same) Adds ROBUSTNESS – need to learn multiple ways to compute the function being learned 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 9

Back to Deep ANNs - Drop Out as an Ensemble Drop Out can be viewed as training an ensemble of ‘thinned’ ANNs - ie, consider all possible ANNs that one can construct by ‘thinning’ the non-output nodes in the original ANN - in each Drop Out step we are training ONE of these (but note that ALL since wgts shared) 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week becomes We implicitly store O (2 N ) networks in 1, where N = # non- output nodes

Warning: At the Research Frontier Research on Deep ANNs changing rapidly, lot of IT-industry money dedicated to it Until recently, people used unsupervised ML to train all the HU layers except the final one (surprisingly, BP works through many levels when much data!) So this ‘slide deck’ likely to be out of date soon, if not already 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 10 11

Neural Network Wrapup ANNs compute weighted sums to make decisions Use (stochastic) gradient descent to adjust weights in order to reduce error (or cost) Only find local minima, though (but good enough!) Impressive testset accuracy, especially Deep ANNs on (mainly) vision tasks and natural language tasks Slow training (GPUs, parallelism, advanced optimization methods, etc help) Learned models hard to interpret 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 1012