Deep Learning Amin Sobhani.

Slides:



Advertisements
Similar presentations
A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Advertisements

Neural networks Introduction Fitting neural networks
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Machine Learning Neural Networks
Neural Networks I CMPUT 466/551 Nilanjan Ray. Outline Projection Pursuit Regression Neural Network –Background –Vanilla Neural Networks –Back-propagation.
Lecture 14 – Neural Networks
Artificial Neural Networks ECE 398BD Instructor: Shobha Vasudevan.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
NEURAL NETWORKS FOR DATA MINING
Classification / Regression Neural Networks 2
Multi-Layer Perceptron
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
EEE502 Pattern Recognition
Neural Network and Deep Learning 王强昌 MLA lab.
Neural Networks 2nd Edition Simon Haykin
Neural Networks Lecture 11: Learning in recurrent networks Geoffrey Hinton.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Big data classification using neural network
Convolutional Neural Network
The Relationship between Deep Learning and Brain Function
Artificial Neural Networks
Randomness in Neural Networks
Data Mining, Neural Network and Genetic Programming
School of Computer Science & Engineering
Recurrent Neural Networks for Natural Language Processing
COMP24111: Machine Learning and Optimisation
Matt Gormley Lecture 16 October 24, 2016
Intro to NLP and Deep Learning
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Classification with Perceptrons Reading:
Intelligent Information System Lab
LECTURE 28: NEURAL NETWORKS
Neural networks (3) Regularization Autoencoder
Deep learning and applications to Natural language processing
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CS 188: Artificial Intelligence
Deep Architectures for Artificial Intelligence
Neuro-Computing Lecture 4 Radial Basis Function Network
network of simple neuron-like computing elements
Artificial Neural Networks
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Neural Networks Geoff Hulten.
Capabilities of Threshold Neurons
Lecture Notes for Chapter 4 Artificial Neural Networks
Neural networks (3) Regularization Autoencoder
COSC 4335: Part2: Other Classification Techniques
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CSC321: Neural Networks Lecture 11: Learning in recurrent networks
Attention for translation
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Introduction to Neural Networks
Image recognition.
PYTHON Deep Learning Prof. Muhammad Saeed.
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Deep Learning Amin Sobhani

Conventional Machine Learning Process natural data in raw form Constructing features by hand Requires domain expertise Difficult and time consuming Raw Input (pixels from an image) Feature Extraction (SIFT, HOG) Learning System (Classifier)

Representation Learning Representation Learning: allows a machine to be fed with raw data and to automatically discover representations needed for detection/classification Deep Learning: representation learning with multiple layers of representation (more than 3) Transformed into higher, slightly more abstract level Very complex functions can be learned

Deep Learning Layers are not handcrafted Features are learned from raw data via a general-purpose learning algorithm https://devblogs.nvidia.com/parallelforall/accelerate-machine-learning-cudnn-deep-neural-network-library/

Applications of Deep Learning Domains in science, business and government Beat current records in image and speech recognition Beaten other machine-learning techniques at Predicting activity of potential drug molecules Analyzing particle accelerator data Reconstructing brain circuits Produced promising results in natural language understanding Topic classification Sentiment analysis Question answering

Overview Supervised Learning Backpropagation to Train Multilayer Architectures Convolution Neural Networks Image Understanding with Deep Convolution Networks Distributed Representation and Language Processing Recurrent Neural Networks Future of Deep Learning

Supervised Learning Most common form of machine learning Data set Labeling Training on data set (tuning parameters, gradient descent) Testing Objective function: measures error between output scores and the desired pattern of scores Modifies internal adjustable parameters (weights) to reduce error

Supervised Learning Objective Function → “Hilly landscape” in high dimensional space of weight values Computes a gradient vector Indicates how much the error would increase or decrease if the weight were increased by a tiny amount Negative gradient vector indicates the direction of steepest descent in this landscape Taking it closer to a minimum, where the output error is low on average

Stochastic Gradient Descent (SGD) Input vector for a few examples Compute the outputs and the errors Compute the average gradient Adjusting the weights Repeated for many small sets from the training set until the average of the objective function stops decreasing Stochastic: small set gives a noisy estimate of the average gradient over all examples

Multi Layer Neural Network Distort the input space to make the classes of data (ex: red and blue lines) linearly separable Illustrative example with only two input units, two hidden units and one output unit

Feed Forward Map a fixed-size input (Ex: an image) to a fixed-size output (Ex: a probability for each of several categories) A set of units compute a weighted sum of their inputs from the previous layer Pass the result through a nonlinear function

Backpropagation to Train Multilayer Architectures Feed Forward the input Calculate the error function Calculate the gradient backward

Convolutional Neural Networks Use many different copies of the same feature detector with different positions Replication greatly reduces the number of features to be learned Enhanced generalization Use several different feature type, each with its own map of replicated detectors.

Convolutional Neural Networks

Image Understanding with Deep Convolutional Networks

Distributed Representation and Language Processing Data is represented as a vector Each element is not mutually dependent Many possible combination for the same input (stochastic representation) Enhanced classification accuracy

Recurrent Neural Networks RNNs process an input sequence one element at a time Tasks that involve sequential input speech, language Trained by backpropagation problematic because the back propagated gradients either grow or shrink at each time step over many time steps they typically explode or vanish

My View Over Future of Deep Learning Expect unsupervised learning to become more important Human and animal learning is largely unsupervised Future progress in vision Systems trained end-to-end Combine ConvNets with RNNS that use reinforcement learning to decide where to look Natural language RNNs systems will become better when they learn strategies for selectively attending to one part at a time Ultimate progress → systems that combine representation learning with complex reasoning

Discussion Deep Learning has already drastically improved the state-of-the-art in image recognition speech recognition natural language understanding Deep Learning requires very little engineering by hand and thus has the potential to be applied to many fields

References Y. LeCun, Y. Bengio, G. Hinton (2015). Deep Learning. Nature 521, 436-444.