Neural Networks for Machine Learning Lecture 1e Three types of learning Geoffrey Hinton with Nitish Srivastava Kevin Swersky.

Slides:



Advertisements
Similar presentations
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
Advertisements

Rerun of machine learning Clustering and pattern recognition.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 24: Non-linear Support Vector Machines Geoffrey Hinton.
Machine learning continued Image source:
Data Visualization STAT 890, STAT 442, CM 462
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Supervised learning Given training examples of inputs and corresponding outputs, produce the “correct” outputs for new inputs Two main scenarios: –Classification:
Introduction to Neural Networks Simon Durrant Quantitative Methods December 15th.
Rise of the Machines Jack Concanon
Neural Networks Lecture 8: Two simple learning algorithms
Lecture 2: Introduction to Machine Learning
Artificial Neural Networks:
This week: overview on pattern recognition (related to machine learning)
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
CSC321: Neural Networks Lecture 13: Learning without a teacher: Autoencoders and Principal Components Analysis Geoffrey Hinton.
CSC321: Neural Networks Lecture 2: Learning with linear neurons Geoffrey Hinton.
CS 445/545 Machine Learning Spring, 2013 See syllabus at
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
Learning from observations
CSC2535: Computation in Neural Networks Lecture 12: Non-linear dimensionality reduction Geoffrey Hinton.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Data Mining and Decision Support
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Neural Networks for Machine Learning Lecture 3a Learning the weights of a linear neuron Geoffrey Hinton with Nitish Srivastava Kevin Swersky.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
CSC2535: Lecture 4: Autoencoders, Free energy, and Minimum Description Length Geoffrey Hinton.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Machine Learning Supervised Learning Classification and Regression
Big data classification using neural network
Unsupervised Learning of Video Representations using LSTMs
Machine learning Image source:
CS 445/545 Machine Learning Winter, 2017
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
Deep Feedforward Networks
Deep Learning Amin Sobhani.
Ananya Das Christman CS311 Fall 2016
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
Lecture 3: Linear Regression (with One Variable)
CS 445/545 Machine Learning Spring, 2017
10701 / Machine Learning.
Estimating Link Signatures with Machine Learning Algorithms
Intelligent Information System Lab
Machine Learning Basics
Machine Learning Ali Ghodsi Department of Statistics
Overview of Supervised Learning
Unsupervised Learning and Neural Networks
Self organizing networks
Unsupervised Learning and Autoencoders
Machine Learning Week 1.
IOT in Transport: Data to Information & Information to Knowledge
Predictive Learning from Data
Collaborative Filtering Matrix Factorization Approach
Dr. Unnikrishnan P.C. Professor, EEE
3.1.1 Introduction to Machine Learning
Artificial Intelligence Lecture No. 28
Deep Learning for Non-Linear Control
Machine Learning Algorithms – An Overview
Multivariate Methods Berlin Chen
Machine learning overview
Neural networks (3) Regularization Autoencoder
Multivariate Methods Berlin Chen, 2005 References:
Machine Learning Support Vector Machine Supervised Learning
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CSC321: Neural Networks Lecture 11: Learning in recurrent networks
Machine Learning – a Probabilistic Perspective
Machine Learning.
What is Artificial Intelligence?
Presentation transcript:

Neural Networks for Machine Learning Lecture 1e Three types of learning Geoffrey Hinton with Nitish Srivastava Kevin Swersky

Types of learning task Supervised learning Learn to predict an output when given an input vector. Reinforcement learning Learn to select an action to maximize payoff. Unsupervised learning Discover a good internal representation of the input.

Two types of supervised learning Each training case consists of an input vector x and a target output t. Regression: The target output is a real number or a whole vector of real numbers. The price of a stock in 6 months time. The temperature at noon tomorrow. Classification: The target output is a class label. The simplest case is a choice between 1 and 0. We can also have multiple alternative labels.

How supervised learning typically works We start by choosing a model-class: A model-class, f, is a way of using some numerical parameters, W, to map each input vector, x, into a predicted output y. Learning usually means adjusting the parameters to reduce the discrepancy between the target output, t, on each training case and the actual output, y, produced by the model. For regression, is often a sensible measure of the discrepancy. For classification there are other measures that are generally more sensible (they also work better).

Reinforcement learning In reinforcement learning, the output is an action or sequence of actions and the only supervisory signal is an occasional scalar reward. The goal in selecting each action is to maximize the expected sum of the future rewards. We usually use a discount factor for delayed rewards so that we don’t have to look too far into the future. Reinforcement learning is difficult: The rewards are typically delayed so its hard to know where we went wrong (or right). A scalar reward does not supply much information.

Unsupervised learning For about 40 years, unsupervised learning was largely ignored by the machine learning community Some widely used definitions of machine learning actually excluded it. Many researchers thought that clustering was the only form of unsupervised learning. It is hard to say what the aim of unsupervised learning is. One major aim is to create an internal representation of the input that is useful for subsequent supervised or reinforcement learning. You can compute the distance to a surface by using the disparity between two images. But you don’t want to learn to compute disparities by stubbing your toe thousands of times.

Other goals for unsupervised learning It provides a compact, low-dimensional representation of the input. High-dimensional inputs typically live on or near a low-dimensional manifold (or several such manifolds). Principal Component Analysis is a widely used linear method for finding a low-dimensional representation. It provides an economical high-dimensional representation of the input in terms of learned features. Binary features are economical. So are real-valued features that are nearly all zero. It finds sensible clusters in the input. This is an example of a very sparse code in which only one of the features is non-zero.