Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015.

Slides:



Advertisements
Similar presentations
Deep Belief Nets and Restricted Boltzmann Machines
Advertisements

Deep Learning Bing-Chen Tsai 1/21.
CIAR Second Summer School Tutorial Lecture 2a Learning a Deep Belief Net Geoffrey Hinton.
CS590M 2008 Fall: Paper Presentation
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Deep Learning and Neural Nets Spring 2015
Stacking RBMs and Auto-encoders for Deep Architectures References:[Bengio, 2009], [Vincent et al., 2008] 2011/03/03 강병곤.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
POSTER TEMPLATE BY: Multi-Sensor Health Diagnosis Using Deep Belief Network Based State Classification Prasanna Tamilselvan.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Deep Learning.
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Network Goodness and its Relation to Probability PDP Class Winter, 2010 January 13, 2010.
Deep Belief Networks for Spam Filtering
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Submitted by:Supervised by: Ankit Bhutani Prof. Amitabha Mukerjee (Y )Prof. K S Venkatesh.
Deep Boltzman machines Paper by : R. Salakhutdinov, G. Hinton Presenter : Roozbeh Gholizadeh.
CIAR Second Summer School Tutorial Lecture 2b Autoencoders & Modeling time series with Boltzmann machines Geoffrey Hinton.
Using Fast Weights to Improve Persistent Contrastive Divergence Tijmen Tieleman Geoffrey Hinton Department of Computer Science, University of Toronto ICML.
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
A shallow introduction to Deep Learning
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
The Boltzmann Machine Psych 419/719 March 1, 2001.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
Geoffrey Hinton CSC2535: 2013 Lecture 5 Deep Boltzmann Machines.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
Training Restricted Boltzmann Machines using Approximations to the Likelihood Gradient Tijmen Tieleman University of Toronto.
CSC2535 Lecture 4 Boltzmann Machines, Sigmoid Belief Nets and Gibbs sampling Geoffrey Hinton.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 18 Learning Boltzmann Machines Geoffrey Hinton.
The Essence of PDP: Local Processing, Global Outcomes PDP Class January 16, 2013.
Constraint Satisfaction and Schemata Psych 205. Goodness of Network States and their Probabilities Goodness of a network state How networks maximize goodness.
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
How to learn a generative model of images Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
CSC321: Introduction to Neural Networks and Machine Learning Lecture 19: Learning Restricted Boltzmann Machines Geoffrey Hinton.
Boltzman Machines Stochastic Hopfield Machines Lectures 11e 1.
Cognitive models for emotion recognition: Big Data and Deep Learning
Chapter 6 Neural Network.
Deep learning Tsai bing-chen 10/22.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
CSC2535 Lecture 5 Sigmoid Belief Nets
CSC2515 Fall 2008 Introduction to Machine Learning Lecture 8 Deep Belief Nets All lecture slides will be available as.ppt,.ps, &.htm at
CSC321: Computation in Neural Networks Lecture 21: Stochastic Hopfield nets and simulated annealing Geoffrey Hinton.
CSC321 Lecture 24 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
Deep Belief Network Training Same greedy layer-wise approach First train lowest RBM (h 0 – h 1 ) using RBM update algorithm (note h 0 is x) Freeze weights.
CSC2535: Computation in Neural Networks Lecture 8: Hopfield nets Geoffrey Hinton.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
CSC321 Lecture 27 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton.
1 Restricted Boltzmann Machines and Applications Pattern Recognition (IC6304) [Presentation Date: ] [ Ph.D Candidate,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
1 Neural Networks MUMT 611 Philippe Zaborowski April 2005.
Some Slides from 2007 NIPS tutorial by Prof. Geoffrey Hinton
Learning Deep Generative Models by Ruslan Salakhutdinov
Energy models and Deep Belief Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Real Neurons Cell structures Cell body Dendrites Axon
Restricted Boltzmann Machines for Classification
Spring Courses CSCI 5922 – Probabilistic Models
Deep Learning Qing LU, Siyuan CAO.
Deep Belief Networks Psychology 209 February 22, 2013.
Structure learning with deep autoencoders
CSSE463: Image Recognition Day 17
network of simple neuron-like computing elements
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
Boltzmann Machine (BM) (§6.4)
CSC321 Winter 2007 Lecture 21: Some Demonstrations of Restricted Boltzmann Machines Geoffrey Hinton.
CSSE463: Image Recognition Day 17
CSSE463: Image Recognition Day 17
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

Stochastic Neural Networks Deep Learning and Neural Nets Spring 2015

Neural Net T-Shirts

A Brief History Of Deterministic And Stochastic Networks 1982 Hopfield Nets Boltzmann Machines / Harmony Nets Back Propagation 2005 Restricted Boltzmann Machines and Deep Belief Nets 1992 Sigmoid Belief Networks 2009 Deep Learning With Back Propagation

Hopfield Networks Binary-threshold units Asynchronous update Symmetric weights Solves an optimization problem  minimize energy (or cost or potential)  maximize harmony (or goodness-of-fit)  search for parameters (activities) that produce the best solution

y h

Hopfield Net As Content Addressible Memory Won’t discuss training procedure because it’s dorky  Hebbian learning Training on set of patterns causes them to become attractors Degraded input is mapped to nearest attractor

Boltzmann Machine Demo Necker Cube Demo (Simon Dennis)

How a Boltzmann machine models data

Three Ways To Specify Inputs Use input to set initial activations  bad idea: initial activations irrelevant once equilibrium is reached Use input to clamp or freeze unit activations  clamped neurons effectively vanish from network and serve as bias on hidden neurons Use input to impose strong bias  set b i such that unit i will (almost) always be off or on

Back To Thermal Equilibrium

no need for back propagation Positive and negative phases  positive phase clamp visible units set hidden randomly run to equilibrium for given T compute expectations +  negative phase set visible and hidden randomly run to equilibrium for T=1 compute expectations -

Why Boltzmann Machine Failed Too slow  loop over training epochs loop over training examples loop over 2 phases (+ and -) loop over annealing schedule for T loop until thermal equilibrium reached loop to sample Sensitivity to annealing schedule Difficulty determining when equilibrium is reached As learning progresses, weights get larger, energy barriers get hard to break -> becomes even slower Back prop was invented shortly after  The need to perform pattern completion wasn’t necessary for most problems (feedforward nets sufficed)

Comments On Boltzmann Machine Learning Algorithm No need for back propagation  reaching thermal equilibrium involves propagating information through network Positive and negative phase  positive phase clamp visible units set hidden randomly run to equilibrium for T=1 compute expectations +  negative phase set visible and hidden randomly run to equilibrium for T=1 compute expectations - Why Boltzmann machine failed (circa 1985)

Restricted Boltzmann Machine Restricted Boltzmann Machine (also known as Harmony Network) Architecture Why positive phase is trivial Contrastive divergence algorithm Example of RBM learning

RBM Generative Model As A Product Of Experts

Deep RBM Autoencoder Hinton & Salakhutdinov (2006)

Deep Belief Nets (DBNs): Using Stacked RBMs As A Generative Model Generative model is not a Boltzmann machine Why do we need symmetric connections between H2 and H3? V H1 H2 H3 H2 H3 V H1

Using A DBN For Supervised Learning 1. Train RBMs in unsupervised fashion 2. In final RBM, include additional units representing class labels 3a. Recognition model  Use feedforward weights and fine tune with back prop 3b. Generative model  Alternating Gibbs sampling between H3 and H4, and feedback weights elsewhere V H1 H2 H3 H4 L H2 H3 V H1 L H4 H2 H3 V H1 L H4

Performance on MNIST (Hinton, Osindero, & Teh, 2006) recognition model generative model