March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU On-Line Handwriting Recognition Transducer device (digitizer) Input: sequence.

Slides:



Advertisements
Similar presentations
Component Analysis (Review)
Advertisements

Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Face Recognition CPSC UTC/CSE.
Face Recognition By Sunny Tang.
Face Recognition and Biometric Systems
Face Recognition Committee Machine Presented by Sunny Tang.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
Combining Shape and Physical Models for Online Cursive Handwriting Synthesis Jue Wang (University of Washington) Chenyu Wu (Carnegie Mellon University)
Face Recognition Under Varying Illumination Erald VUÇINI Vienna University of Technology Muhittin GÖKMEN Istanbul Technical University Eduard GRÖLLER Vienna.
© 2003 by Davi GeigerComputer Vision September 2003 L1.1 Face Recognition Recognized Person Face Recognition.
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Eigenfaces As we discussed last time, we can reduce the computation by dimension reduction using PCA –Suppose we have a set of N images and there are c.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Implementing a reliable neuro-classifier
PCA Channel Student: Fangming JI u Supervisor: Professor Tom Geoden.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Ch. 10: Linear Discriminant Analysis (LDA) based on slides from
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Comparing Kernel-based Learning Methods for Face Recognition Zhiguo Li
Face Recognition: An Introduction
Foundation of High-Dimensional Data Visualization
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
(Off-Line) Cursive Word Recognition Tal Steinherz Tel-Aviv University.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
PCA & LDA for Face Recognition
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Face Recognition and Feature Subspaces
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Loop Investigation for Cursive Handwriting Processing and Recognition By Tal Steinherz Advanced Seminar (Spring 05)
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Using Support Vector Machines to Enhance the Performance of Bayesian Face Recognition IEEE Transaction on Information Forensics and Security Zhifeng Li,
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
Face Recognition: An Introduction
Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Presented by Xianwang Wang Masashi Sugiyama.
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Discriminant Analysis
Feature extraction using fuzzy complete linear discriminant analysis The reporter : Cui Yan
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Dimensionality reduction
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
June 25-29, 2006ICML2006, Pittsburgh, USA Local Fisher Discriminant Analysis for Supervised Dimensionality Reduction Masashi Sugiyama Tokyo Institute of.
Principal Component Analysis and Linear Discriminant Analysis for Feature Reduction Jieping Ye Department of Computer Science and Engineering Arizona State.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
3D Face Recognition Using Range Images Literature Survey Joonsoo Lee 3/10/05.
Feature Extraction 主講人:虞台文.
1 An introduction to support vector machine (SVM) Advisor : Dr.Hsu Graduate : Ching –Wen Hong.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Arabic Handwriting Recognition Thomas Taylor. Roadmap  Introduction to Handwriting Recognition  Introduction to Arabic Language  Challenges of Recognition.
Martina Uray Heinz Mayer Joanneum Research Graz Institute of Digital Image Processing Horst Bischof Graz University of Technology Institute for Computer.
Combining Neural Networks and Context-Driven Search for On- Line, Printed Handwriting Recognition in the Newton Larry S. Yaeger, Brandn J. Web, and Richard.
Principal Component Analysis (PCA)
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Face Recognition and Feature Subspaces
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Brief Review of Recognition + Context
Principal Component Analysis
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Paper Reading Dalong Du April.08, 2011.
EE 492 ENGINEERING PROJECT
Presentation transcript:

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU On-Line Handwriting Recognition Transducer device (digitizer) Input: sequence of point coordinates with pen-down/up signals from the digitizer Stroke: sequence of points from pen-down to pen-up signals Word: sequence of one or more strokes.

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU System Overview Input Pre-processing (high curvature points) Segmentation Recognition Engine Dictionary Character Recognizer Context Models Word Candidates

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Segmentation Hypotheses High-curvature points and segmentation points:

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Character Recognition I Fisher Discriminant Analysis (FDA): improves over PCA (Principal Component Analysis). Original spaceProjection space Linear projec- tion p=W T x Training set: 1040 lowercase letters, Test set: 520 lowercase letters Test results: 91.5% correct

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Fisher Discriminant Analysis Between-class scatter matrix –C: number of classes –N i : number of data vectors in class i –  i : mean vector of class i and  : mean vector Within-class scatter matrix –v j i : j-th data vector of class i.

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Given a projection matrix W (of size n by m) and its linear transformation, the between-class scatter in the projection space is Similarly

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Optimization formulation of the fisher projection solution: (  B,  W are scatter matrices in projection space) Fisher Discriminant Analysis (cont.)

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU FDA (continued) Construction of the Fisher projection matrix: –Compute the n eigenvalues and eigenvectors of the generalized eigenvalue problem: –Retain the m eigenvectors having the largest eigenvalues. They form the columns of the target projection matrix.

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Character Recognition Results Training set: 1040 lowercase letters Test set: 520 lowercase letters Test results:

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Challenge I The problem of the previous approach is: non-characters are classified as characters. When applied to cursive words it creates several/too many non-sense word hypothesis by extracting characters where they don’t seem to exist. More generally, one wants to be able to generate shapes and their deformations.

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Challenge II How to extract reliable local geometric features of images (corners, contour tangents, contour curvature, …) ? How to group them ? Large size data base to match one input, how to do it fast ? Hierarchical clustering of the database, possibly over a tree structure or some general graph. How to do it ? Which criteria to cluster ? Which methods to use it ?

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Recognition Engine Integrates all available information, generates and grows the word-level hypotheses. Most general form: graph and its search. Hypothesis Propagation Network

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Hypothesis Propagation Network H (t, m) Class m's legal predecessors "a""b”"z""y" List length Time T t Look- back window range m Recognition of 85% on 100 words (not good)

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Challenge III How to search more efficiently in this network and more generally on Bayesian networks ?

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU Visual Bigram Models (VBM) Some characters can be very ambiguous when isolated: “9” and “g”; “e” and “l”; “o” and “0”; etc, but more obvious when put in a context. Character heights Relative height ratio and positioning “go” “90”

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU VBM: Parameters h1h1 h2h2 h top 1 top 2 bot 1 bot 2 Height Diff. Ratio: HDR = (h 1 - h 2 ) / h Top Diff. Ratio: TDR = (top 1 - top 2 ) / h Bottom Diff. Ratio: BDR = (bot 1 - bot 2 ) / h

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU VBM: Ascendancy Categories Total 9 visual bigram categories (instead of 26x26=676).

March 15-17, 2002Work with student Jong Oh Davi Geiger, Courant Institute, NYU VBM: Test Results