OUTLINE Course description, What is pattern recognition, Cost of error, Decision boundaries, The desgin cycle.

Slides:



Advertisements
Similar presentations
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
Advertisements

ECE 8443 – Pattern Recognition Objectives: Course Introduction Typical Applications Resources: Syllabus Internet Books and Notes D.H.S: Chapter 1 Glossary.

Chapter 1: Introduction to Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition 9/23/2008 Instructor: Wen-Hung Liao, Ph.D.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Machine Vision and Dig. Image Analysis 1 Prof. Heikki Kälviäinen C50A6100 Lectures 12: Object Recognition Professor Heikki Kälviäinen Machine Vision and.
Pattern Classification, Chapter 1 1 Basic Probability.
Lecture #1COMP 527 Pattern Recognition1 Pattern Recognition Why? To provide machines with perception & cognition capabilities so that they could interact.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1. Introduction to Pattern Recognition and Machine Learning. Prof. A.L. Yuille. Dept. Statistics. UCLA. Stat 231. Fall 2004.
Lecture 1: Introduction to Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture Notes by Neşe Yalabık Spring 2011.
CEN 592 PATTERN RECOGNITION Spring Term CEN 592 PATTERN RECOGNITION Spring Term DEPARTMENT of INFORMATION TECHNOLOGIES Assoc. Prof.
Introduction Mohammad Beigi Department of Biomedical Engineering Isfahan University
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Introduction to Pattern Recognition Charles Tappert Seidenberg School of CSIS, Pace University.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
: Chapter 1: Introduction 1 Montri Karnjanadecha ac.th/~montri Principles of Pattern Recognition.
This week: overview on pattern recognition (related to machine learning)
TINONS1 Nonlinear SP and Pattern recognition
Institute of Systems and Robotics ISR – Coimbra Mobile Robotics Lab Bayesian Approaches 1 1 jrett.
Chapter 4 CONCEPTS OF LEARNING, CLASSIFICATION AND REGRESSION Cios / Pedrycz / Swiniarski / Kurgan.
Lecture 2: Bayesian Decision Theory 1. Diagram and formulation
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
IBS-09-SL RM 501 – Ranjit Goswami 1 Basic Probability.
Perception Introduction Pattern Recognition Image Formation
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
2. Bayes Decision Theory Prof. A.L. Yuille Stat 231. Fall 2004.
1 Pattern Recognition Concepts How should objects be represented? Algorithms for recognition/matching * nearest neighbors * decision tree * decision functions.
Compiled By: Raj G Tiwari.  A pattern is an object, process or event that can be given a name.  A pattern class (or category) is a set of patterns sharing.
Nearest Neighbor (NN) Rule & k-Nearest Neighbor (k-NN) Rule Non-parametric : Can be used with arbitrary distributions, No need to assume that the form.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
1 E. Fatemizadeh Statistical Pattern Recognition.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall 2004 AdaBoost.. Binary Classification. Read 9.5 Duda,
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Introduction to Pattern Recognition (การรู้จํารูปแบบเบื้องต้น)
Pattern Recognition NTUEE 高奕豪 2005/4/14. Outline Introduction Definition, Examples, Related Fields, System, and Design Approaches Bayesian, Hidden Markov.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
Objectives: Loss Functions Risk Min. Error Rate Class. Resources: DHS – Chap. 2 (Part 1) DHS – Chap. 2 (Part 2) RGO - Intro to PR MCE for Speech MCE for.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
Artificial Intelligence
Machine Learning for Computer Security
Pattern Classification Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 Dr. Ding Yuxin Pattern Recognition.
Introduction Machine Learning 14/02/2017.
Special Topics In Scientific Computing
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
School of Computer Science & Engineering
LECTURE 01: COURSE OVERVIEW
CS668: Pattern Recognition Ch 1: Introduction
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
What is Pattern Recognition?
Introduction to Pattern Recognition and Machine Learning
Introduction to Pattern Recognition
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
An Introduction to Supervised Learning
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
LECTURE 01: COURSE OVERVIEW
LECTURE 23: INFORMATION THEORY REVIEW
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Presentation transcript:

OUTLINE Course description, What is pattern recognition, Cost of error, Decision boundaries, The desgin cycle

The objective of the course is to make student familiar with general approaches such as – Bayes classification, – discriminant functions, – decision trees, – nearest neighbor rule, – neural networks for pattern recognition. Course description

Course Book: – R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, New York: John Wiley, 2001, Course description

WEEKLY SCHEDULE AND PRE-STUDY PAGES Wee k TopicsPre-study Pages 1IntroductionChapter 1 (main text) 2Bayesian Decision TheoryChapter 2 3Bayesian Decision TheoryChapter 2 4Bayesian Decision TheoryChapter 2 5Maximum – Likelihood and Bayesian Parameter EstimationChapter 3 6Maximum – Likelihood and Bayesian Parameter EstimationChapter 3 7Nonparametric TechniquesChapter 4 8Nonparametric TechniquesChapter 4 9Linear Discriminant FunctionsChapter 5 10Linear Discriminant FunctionsChapter 5 11Multilayer Neural NetworksChapter 6 12Nonmetric MethodsChapter 8 13Unsupervised Learning and ClusteringChapter 10 14Unsupervised Learning and ClusteringChapter 10

EVALUATION SYSTEM IN-TERM STUDIESQUANTITYPERCENTAGE Mid-terms22x20=40 Assignment330 Final Exam130 TOTAL100 Course description

What is a pattern ? an entity, vaguely defined, that could be given a name, e.g.: – fingerprint image, – handwritten word, – human face, – speech signal, – DNA sequence,

What is pattern recognition ? a discipline which learn some theories and methods to design machines that can recognize patterns in noisy data or complex environment (Srihari,Govindaraju).

What is pattern recognition ? a scientific discipline whose aim is the classification of the objects into a lot of categories or classes. Pattern recognition is also a integral part in most machine intelligence system built for decision making (Sergios Theodoridis).

What is pattern recognition ? the act of taking in raw data and making an action based on the category of the pattern (Duda and Hart)

What is pattern recognition ? Pattern recognition is the study of how machines can: – observe the environment, – learn to distinguish patterns of interest, – make sound and reasonable decisions about the categories of the patterns.

What is pattern recognition ? Some Applications:

What is pattern recognition ? Some Applications:

What is pattern recognition ? Some Applications:

What is pattern recognition ? Some Applications:

What is pattern recognition ? Some Applications:

What is pattern recognition ? Some Applications:

An Example Suppose that: – A fish packing plant wants to automate the process of sorting incoming fish on a conveyor belt according to species, – There are two species: Sea bass, Salmon.

An Example How to distinguish one specie from the other ? (length, width, weight, number and shape of fins, tail shape,etc.)

An Example Suppose somebody at the fish plant say us that: – Sea bass is generally longer than a salmon Then our models for the fish: – Sea bass have some typical length, and this is greater than that for salmon.

An Example Then length becomes a feature, We might attempt to classify the fish by seeing whether or not the length of a fish exceeds some critical value (threshold value) l*.

An Example How to desie on the critical value (threshold value) ?

An Example How to desie on the critical value (threshold value) ? – We could obtain some training samples of different types of fish, – make length measurements, – Inspect the results.

An Example Measurement results on the training sample related to two species.

An Example Can we reliably seperate sea bass from salmon by using length as a feature ? Remember our model: – Sea bass have some typical length, and this is greater than that for salmon.

An Example From histogram we can see that single criteria is quite poor.

An Example It is obvious that length is not a good feature. What we can do to seperate sea bass from salmon?

An Example What we can do to seperate sea bass from salmon? Try another feature: – average lightness of the fish scales.

An Example Can we reliably seperate sea bass from salmon by using lightness as a feature ?

An Example Lighness is better than length as a feature but again there are some problems.

An Example Suppose we also know that: – Sea bass are typically wider than salmon. We can use more than one feature for our decision: – Lightness (x 1 ) and width (x 2 )

An Example Each fish is now a point in two dimension. – Lightness (x 1 ) and width (x 2 )

An Example Each fish is now a point in two dimension. – Lightness (x 1 ) and width (x 2 )

An Example Each fish is now a point in two dimension. – Lightness (x 1 ) and width (x 2 )

Cost of error Cost of different errors must be considered when making decisions, We try to make a decision rule so as to minimize such a cost, This is the central task of decision theory.

Cost of error For example, if the fish packing company knows that: – Customers who buy salmon will object if they see sea bass in their cans. – Customers who buy sea bass will not be unhappy if they occasionally see some expensive salmon in their cans.

Decision boundaries We can perform better if we use more complex decision boundaries.

Decision boundaries There is a trade of between complexity of the decision rules and their performances to unknown samples. Generalization: The ability of the classifier to produce correct results on novel patterns. Simplify the decision boundary!

The design cycle

Collect data: – Collect train and test data Choose features: – Domain dependence and prior information, – Computational cost and feasibility, – Discriminative features, – Invariant features with respect to translation, rotation and scale, – Robust features with respect to occlusion, distortion, deformation, and variations in environment.

The design cycle Choose model: – Types of models: templates, decision-theoretic or statistical, syntactic or structural, neural, and hybrid. Train classifier: – learn the rule from data, Evaluate classifier: – estimate the performance – problems of overfitting and generalization.

References Srihari, S.N., Covindaraju, Pattern recognition, Chapman &Hall, London, , 1993, Sergios Theodoridis, Konstantinos Koutroumbas, pattern recognition, Pattern Recognition,Elsevier(USA)),1982 R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, New York: John Wiley, 2001,