Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.

Slides:



Advertisements
Similar presentations
Introduction to Support Vector Machines (SVM)
Advertisements

Face Recognition: A Convolutional Neural Network Approach
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Thesis title: “Studies in Pattern Classification – Biological Modeling, Uncertainty Reasoning, and Statistical Learning” 3 parts: (1)Handwritten Digit.
Algorithm-Independent Machine Learning Anna Egorova-Förster University of Lugano Pattern Classification Reading Group, January 2007 All materials in these.
Lecture 14 – Neural Networks
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
Learning From Data Chichang Jou Tamkang University.
Support Vector Machines Kernel Machines
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
ENN: Extended Nearest Neighbor Method for Pattern Recognition
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Image Classification 영상분류
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
An Introduction to Support Vector Machines (M. Law)
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
1 E. Fatemizadeh Statistical Pattern Recognition.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Handwritten digit recognition
Face Detection Using Large Margin Classifiers Ming-Hsuan Yang Dan Roth Narendra Ahuja Presented by Kiang “Sean” Zhou Beckman Institute University of Illinois.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Analysis of Classification Algorithms In Handwritten Digit Recognition Logan Helms Jon Daniele.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
1 Convolutional neural networks Abin - Roozgard. 2  Introduction  Drawbacks of previous neural networks  Convolutional neural networks  LeNet 5 
PatReco: Introduction Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Today’s Lecture Neural networks Training
CS 9633 Machine Learning Support Vector Machines
Convolutional Neural Network
CEE 6410 Water Resources Systems Analysis
Deep Learning Amin Sobhani.
Non-Parameter Estimation
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Learning Mid-Level Features For Recognition
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Machine Learning Basics
Recognition using Nearest Neighbor (or kNN)
Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan.
Bias and Variance of the Estimator
Non-linear classifiers Neural networks
Recognition - III.
State-of-the-art face recognition systems
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
Outline Parameter estimation – continued Non-parametric methods.
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Learning with information of features
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
Convolutional neural networks Abin - Roozgard.
Creating Data Representations
Biointelligence Laboratory, Seoul National University
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Using Manifold Structure for Partially Labeled Classification
Face Recognition: A Convolutional Neural Network Approach
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Presented by Xu Miao April 20, 2005
Derek Hoiem CS 598, Spring 2009 Jan 27, 2009
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Image recognition.
Presentation transcript:

Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, November, 1998.

Invariant Object Recognition The central goal of computer vision research is to detect and recognize objects invariant to scale, viewpoint, illumination, and other changes November 21, 2018 Computer Vision

(Invariant) Object Recognition November 21, 2018 Computer Vision

Generalization Performance Many classifiers are available Maximum likelihood estimation, Bayesian estimation, Parzen Windows, Kn-nearest neighbor, discriminant functions, support vector machines, neural networks, decision trees, ....... Which method is the best to classify unseen test data? The performance is often determined by features In addition, we are interested in systems that can solve a particular problem well November 21, 2018 Computer Vision

Error Rate on Hand Written Digit Recognition November 21, 2018 Computer Vision

No Free Lunch Theorem November 21, 2018 Computer Vision

No Free Lunch Theorem – cont. November 21, 2018 Computer Vision

Ugly Duckling Theorem In the absence of prior information, there is no principled reason to prefer one representation over another. November 21, 2018 Computer Vision

Bias and Variance Dilemma Regression Find an estimate of a true but unknown function F(x) based on n samples generated by F(x) Bias – the difference between the expected value and the true value; a low bias means on average we will accurately estimate F from D Variance – the variability of estimation; a low bias means that the estimate does not change much as the training set varies. November 21, 2018 Computer Vision

Bias-Variance Dilemma When the training data is finite, there is an intrinsic problem of any classifier function If the function is very generic, i.e., a non-parametric family, it suffers from high variance If the function is very specific, i.e., a parametric family, it suffers from high bias The central problem is to design a family of classifiers a priori such that both the variance and bias are low November 21, 2018 Computer Vision

November 21, 2018 Computer Vision

Bias and Variance vs. Model Complexity November 21, 2018 Computer Vision

Gap Between Training and Test Error Typically the performance of a classifier on a disjoint test set will be larger than that on the training set Where P is the number of training examples, h a measure of capacity (model complexity), a between 0.5 and 1, and k a constant November 21, 2018 Computer Vision

Check Reading System November 21, 2018 Computer Vision

End-to-End Training November 21, 2018 Computer Vision

Graph Transformer Networks November 21, 2018 Computer Vision

Training Using Gradient-Based Learning A multiple module system can be trained using a gradient-based method Similar to backpropagation used for multiple layer perceptrons November 21, 2018 Computer Vision

Convolutional Networks November 21, 2018 Computer Vision

Handwritten Digit Recognition Using a Convolutional Network November 21, 2018 Computer Vision

Training a Convolutional Network The loss function used is Training algorithm is stochastic diagonal Levenberg-Marquardt RBF output is given by November 21, 2018 Computer Vision

MNIST Dataset 60,000 training images 10,000 test images There are several different versions of the dataset November 21, 2018 Computer Vision

Experimental Results November 21, 2018 Computer Vision

Experimental Results November 21, 2018 Computer Vision

Distorted Patterns By using distorted patterns, the training error dropped to 0.8% from 0.95% without deformation November 21, 2018 Computer Vision

Misclassified Examples November 21, 2018 Computer Vision

Comparison November 21, 2018 Computer Vision

Rejection Performance November 21, 2018 Computer Vision

Number of Operations Unit: Thousand operations November 21, 2018 Computer Vision

Memory Requirements November 21, 2018 Computer Vision

Robustness November 21, 2018 Computer Vision

Convolutional Network for Object Recognition November 21, 2018 Computer Vision

NORB Dataset November 21, 2018 Computer Vision

Convolutional Network for Object Recognition November 21, 2018 Computer Vision

Experimental Results November 21, 2018 Computer Vision

Jittered Cluttered Dataset November 21, 2018 Computer Vision

Experimental Results November 21, 2018 Computer Vision

Face Detection November 21, 2018 Computer Vision

Face Detection November 21, 2018 Computer Vision

Multiple Object Recognition Based on heuristic over segmentation It avoids making hard decisions about segmentation by taking a large number of different segmentations November 21, 2018 Computer Vision

Graph Transformer Network for Character Recognition November 21, 2018 Computer Vision

Recognition Transformer and Interpretation Graph November 21, 2018 Computer Vision

Viterbi Training November 21, 2018 Computer Vision

Discriminative Viterbi Training

Discriminative Forward Training November 21, 2018 Computer Vision

Space Displacement Neural Networks By considering all possible locations, one can avoid explicit segmentation Similar to detection and recognition November 21, 2018 Computer Vision

Space Displacement Neural Networks We can replicate convolutional networks at all possible locations November 21, 2018 Computer Vision

Space Displacement Neural Networks November 21, 2018 Computer Vision

Space Displacement Neural Networks November 21, 2018 Computer Vision

Space Displacement Neural Networks November 21, 2018 Computer Vision

SDNN/HMM System November 21, 2018 Computer Vision

Graph Transformer Networks and Transducers November 21, 2018 Computer Vision

On-line Handwriting Recognition System November 21, 2018 Computer Vision

On-line Handwriting Recognition System November 21, 2018 Computer Vision

Comparative Results November 21, 2018 Computer Vision

Check Reading System November 21, 2018 Computer Vision

Confidence Estimation November 21, 2018 Computer Vision

Summary By carefully designing systems with desired invariance properties, one can often achieve better generalization performance by limiting system’s capacity Multiple module systems can be trained often effectively using gradient-based learning methods Even though in theory local gradient-based methods are subject to local minima, in practice it seems it is not a serious problem Incorporating contextual information into recognition systems are often critical for real world applications End-to-end training is often more effective November 21, 2018 Computer Vision