Introduction to Pattern Recognition

Slides:



Advertisements
Similar presentations
1 Image Classification MSc Image Processing Assignment March 2003.
Advertisements

Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
An Overview of Machine Learning
Supervised Learning Recap
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
The Nature of Statistical Learning Theory by V. Vapnik
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Chapter 1: Introduction to Pattern Recognition
Lecture 20 Object recognition I
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition 9/23/2008 Instructor: Wen-Hung Liao, Ph.D.
Chapter 2: Pattern Recognition
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Measuring Model Complexity (Textbook, Sections ) CS 410/510 Thurs. April 27, 2007 Given two hypotheses (models) that correctly classify the training.
Machine Vision and Dig. Image Analysis 1 Prof. Heikki Kälviäinen C50A6100 Lectures 12: Object Recognition Professor Heikki Kälviäinen Machine Vision and.
Classification and application in Remote Sensing.
Lecture #1COMP 527 Pattern Recognition1 Pattern Recognition Why? To provide machines with perception & cognition capabilities so that they could interact.
Visual Recognition Tutorial
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Statistical Learning: Pattern Classification, Prediction, and Control Peter Bartlett August 2002, UC Berkeley CIS.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Introduction to machine learning
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture Notes by Neşe Yalabık Spring 2011.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
CEN 592 PATTERN RECOGNITION Spring Term CEN 592 PATTERN RECOGNITION Spring Term DEPARTMENT of INFORMATION TECHNOLOGIES Assoc. Prof.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Introduction Mohammad Beigi Department of Biomedical Engineering Isfahan University
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
: Chapter 1: Introduction 1 Montri Karnjanadecha ac.th/~montri Principles of Pattern Recognition.
TINONS1 Nonlinear SP and Pattern recognition
Data Mining Joyeeta Dutta-Moscato July 10, Wherever we have large amounts of data, we have the need for building systems capable of learning information.
Mehdi Ghayoumi Kent State University Computer Science Department Summer 2015 Exposition on Cyber Infrastructure and Big Data.
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
2. Bayes Decision Theory Prof. A.L. Yuille Stat 231. Fall 2004.
Compiled By: Raj G Tiwari.  A pattern is an object, process or event that can be given a name.  A pattern class (or category) is a set of patterns sharing.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory Pattern Recognition Prof.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Introduction to Pattern Recognition. Pattern Recognition.
An Introduction to Support Vector Machine (SVM)
Signature Verification
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Data Mining and Decision Support
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory 1 Prof. George Papadourakis,
Machine Learning Lecture 1: Intro + Decision Trees Moshe Koppel Slides adapted from Tom Mitchell and from Dan Roth.
Pattern Recognition NTUEE 高奕豪 2005/4/14. Outline Introduction Definition, Examples, Related Fields, System, and Design Approaches Bayesian, Hidden Markov.
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
PatReco: Introduction Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Introduction to Pattern Recognition Kama Jambi
CS 9633 Machine Learning Support Vector Machines
Machine Learning for Computer Security
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
CH. 1: Introduction 1.1 What is Machine Learning Example:
제 4 장 패턴인식 이해.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Presentation transcript:

Introduction to Pattern Recognition Center for Machine Perception Czech Technical University in Prague Vojtěch Franc xfrancv@cmp.felk.cvut.cz

What is pattern recognition? “The assignment of a physical object or event to one of several prespecified categeries” -- Duda & Hart A pattern is an object, process or event that can be given a name. A pattern class (or category) is a set of patterns sharing common attributes and usually originating from the same source. During recognition (or classification) given objects are assigned to prescribed classes. A classifier is a machine which performs classification.

Examples of applications Handwritten: sorting letters by postal code, input device for PDA‘s. Printed texts: reading machines for blind people, digitalization of text documents. Optical Character Recognition (OCR) Biometrics Diagnostic systems Military applications Face recognition, verification, retrieval. Finger prints recognition. Speech recognition. Medical diagnosis: X-Ray, EKG analysis. Machine diagnostics, waster detection. Automated Target Recognition (ATR). Image segmentation and analysis (recognition from aerial or satelite photographs).

Approaches Statistical PR: based on underlying statistical model of patterns and pattern classes. Structural (or syntactic) PR: pattern classes represented by means of formal structures as grammars, automata, strings, etc. Neural networks: classifier is represented as a network of cells modeling neurons of the human brain (connectionist approach).

Basic concepts Feature vector - is a point in feature space . Pattern Feature vector - A vector of observations (measurements). - is a point in feature space . Hidden state - Cannot be directly measured. - Patterns with equal hidden state belong to the same class. Task - To design a classifer (decision rule) which decides about a hidden state based on an onbservation.

Example height Task: jockey-hoopster recognition. The set of hidden state is The feature space is weight Training examples Linear classifier:

Components of PR system Sensors and preprocessing Feature extraction Class assignment Pattern Classifier Learning algorithm Teacher Sensors and preprocessing. A feature extraction aims to create discriminative features good for classification. A classifier. A teacher provides information about hidden state -- supervised learning. A learning algorithm sets PR from training examples.

Feature extraction Task: to extract features which are good for classification. Good features: Objects from the same class have similar feature values. Objects from different classes have different values. “Good” features “Bad” features

Feature extraction methods Feature selection Problem can be expressed as optimization of parameters of featrure extractor . Supervised methods: objective function is a criterion of separability (discriminability) of labeled examples, e.g., linear discriminat analysis (LDA). Unsupervised methods: lower dimesional representation which preserves important characteristics of input data is sought for, e.g., principal component analysis (PCA).

Classifier A classifier partitions feature space X into class-labeled regions such that and The classification consists of determining to which region a feature vector x belongs to. Borders between decision boundaries are called decision regions.

Representation of classifier A classifier is typically represented as a set of discriminant functions The classifier assigns a feature vector x to the i-the class if Feature vector Class identifier Discriminant function

Bayesian decision making The Bayesian decision making is a fundamental statistical approach which allows to design the optimal classifier if complete statistical model is known. Definition: Obsevations Hidden states Decisions A loss function A decision rule A joint probability Task: to design decision rule q which minimizes Bayesian risk

Example of Bayesian task Task: minimization of classification error. A set of decisions D is the same as set of hidden states Y. 0/1 - loss function used The Bayesian risk R(q) corresponds to probability of misclassification. The solution of Bayesian task is

Limitations of Bayesian approach The statistical model p(x,y) is mostly not known therefore learning must be employed to estimate p(x,y) from training examples {(x1,y1),…,(x,y)} -- plug-in Bayes. Non-Bayesian methods offers further task formulations: A partial statistical model is avaliable only: p(y) is not known or does not exist. p(x|y,) is influenced by a non-random intervetion . The loss function is not defined. Examples: Neyman-Pearson‘s task, Minimax task, etc.

Discriminative approaches Given a class of classification rules q(x;θ) parametrized by θ the task is to find the “best” parameter θ* based on a set of training examples {(x1,y1),…,(x,y)} -- supervised learning. The task of learning: recognition which classification rule is to be used. The way how to perform the learning is determined by a selected inductive principle.

Empirical risk minimization principle The true expected risk R(q) is approximated by empirical risk with respect to a given labeled training set {(x1,y1),…,(x,y)}. The learning based on the empirical minimization principle is defined as Examples of algorithms: Perceptron, Back-propagation, etc.

Overfitting and underfitting Problem: how rich class of classifications q(x;θ) to use. underfitting good fit overfitting Problem of generalization: a small emprical risk Remp does not imply small true expected risk R.

Structural risk minimization principle Statistical learning theory -- Vapnik & Chervonenkis. An upper bound on the expected risk of a classification rule qQ where  is number of training examples, h is VC-dimension of class of functions Q and 1- is confidence of the upper bound. SRM principle: from a given nested function classes Q1,Q2,…,Qm, such that select a rule q* which minimizes the upper bound on the expected risk.

Unsupervised learning Input: training examples {x1,…,x} without information about the hidden state. Clustering: goal is to find clusters of data sharing similar properties. A broad class of unsupervised learning algorithms: Classifier Classifier Learning algorithm Learning algorithm (supervised)

Example of unsupervised learning algorithm k-Means clustering: Goal is to minimize Classifier Learning algorithm

References Books Journals Duda, Heart: Pattern Classification and Scene Analysis. J. Wiley & Sons, New York, 1982. (2nd edition 2000). Fukunaga: Introduction to Statistical Pattern Recognition. Academic Press, 1990. Bishop: Neural Networks for Pattern Recognition. Claredon Press, Oxford, 1997. Schlesinger, Hlaváč: Ten lectures on statistical and structural pattern recognition. Kluwer Academic Publisher, 2002. Journals Journal of Pattern Recognition Society. IEEE transactions on Neural Networks. Pattern Recognition and Machine Learning.