BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Face Recognition Sumitha Balasuriya.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
1 CPC group SeminarThursday, June 1, 2006 Classification techniques for Hand-Written Digit Recognition Venkat Raghavan N. S., Saneej B. C., and Karteek.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Machine Learning Lecture 8 Data Processing and Representation
IDEAL 2005, 6-8 July, Brisbane Multiresolution Analysis of Connectivity Atul Sajjanhar, Deakin University, Australia Guojun Lu, Monash University, Australia.
Computer Vision – Image Representation (Histograms)
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Processing Digital Images. Filtering Analysis –Recognition Transmission.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Distinguishing Photographic Images and Photorealistic Computer Graphics Using Visual Vocabulary on Local Image Edges Rong Zhang,Rand-Ding Wang, and Tian-Tsong.
Face Recognition Jeremy Wyatt.
CSE 300: Software Reliability Engineering Topics covered: Software metrics and software reliability.
Exploring Microarray data Javier Cabrera. Outline 1.Exploratory Analysis Steps. 2.Microarray Data as Multivariate Data. 3.Dimension Reduction 4.Correlation.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Image Segmentation by Clustering using Moments by, Dhiraj Sakumalla.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Chapter 3 Data Exploration and Dimension Reduction 1.
Hyperspectral Imaging Alex Chen 1, Meiching Fong 1, Zhong Hu 1, Andrea Bertozzi 1, Jean-Michel Morel 2 1 Department of Mathematics, UCLA 2 ENS Cachan,
Presented by Tienwei Tsai July, 2005
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Face Recognition: An Introduction
Texture Detection & Texture related clustering C601 Project Jing Qin Fall 2003.
Descriptive Statistics vs. Factor Analysis Descriptive statistics will inform on the prevalence of a phenomenon, among a given population, captured by.
CSE 185 Introduction to Computer Vision Face Recognition.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
CSSE463: Image Recognition Day 27 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Image Segmentation by Histogram Thresholding Venugopal Rajagopal CIS 581 Instructor: Longin Jan Latecki.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Feature Extraction 主講人:虞台文.
Next, this study employed SVM to classify the emotion label for each EEG segment. The basic idea is to project input data onto a higher dimensional feature.
Presented by: Muhammad Wasif Laeeq (BSIT07-1) Muhammad Aatif Aneeq (BSIT07-15) Shah Rukh (BSIT07-22) Mudasir Abbas (BSIT07-34) Ahmad Mushtaq (BSIT07-45)
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
CSSE463: Image Recognition Day 25 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Visible-Surface Detection Methods. To identify those parts of a scene that are visible from a chosen viewing position. Surfaces which are obscured by.
Principal Components Analysis ( PCA)
Unsupervised Learning II Feature Extraction
1. 2 What is Digital Image Processing? The term image refers to a two-dimensional light intensity function f(x,y), where x and y denote spatial(plane)
Principal Component Analysis (PCA)
Dimensionality Reduction
Exploring Microarray data
University of Ioannina
LECTURE 10: DISCRIMINANT ANALYSIS
Recognition: Face Recognition
Principal Component Analysis (PCA)
Principal Component Analysis
Outline Announcement Texture modeling - continued Some remarks
Historic Document Image De-Noising using Principal Component Analysis (PCA) and Local Pixel Grouping (LPG) Han-Yang Tang1, Azah Kamilah Muda1, Yun-Huoy.
CSSE463: Image Recognition Day 25
Feature space tansformation methods
LECTURE 09: DISCRIMINANT ANALYSIS
Principal Component Analysis
第 九 章 影像邊緣偵測 9-.
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004

OUTLINE Introduction Background Learning and Letter Detection What is “Texture”? PCA based texture representation Texture detection approach Results Analysis and Conclusions References

INTRODUCTION Document image understanding involves: Layout segmentation Logical labeling of blocks at different levels. Simplest : Text/Non-text separation Knowledge of further information such as type style should be useful in many applications e.g.logical layout recognition, document image indexing and retrieval. Proposed generic method based on visual perception “TEXTURE”

What is “Texture”? Texture can be defined as that where “there is a significant variation in intensity levels between nearby pixels; that is, at the limit of resolution, there is non-homogeneity. An Important approach for describing a region is to quantify its texture content

APPLICATIONS OF “TEXTURE” Simplest use : Physical segmentation by classification of blocks using 2 or 3 classes (text/non-text, text/image/line drawing) by simple features like “black/white transitions” Further analysis of document structuring: Characterizing of fonts (using geometrical properties, statistical features or generic techniques like Feature Based Interaction Maps (FBIM)), skew-detection.

APPLICATIONS OF “TEXTURE” ( contd…) Application specific to this project : Background learning and Letter detection using “Texture” with “Principal Component Analysis” (PCA). This document analysis is a necessary pre- processing stage for many document-processing systems such as : – Optical Character Recognition. (OCR) – Document Retrieval. – Document Compression.

PRINCIPAL COMPONENT ANALYSIS (PCA) Technique capable of deriving low dimensional representation which is applied extensively to identify texture of images. Involves a mathematical procedure that transforms a number of possibly correlated variables into a smaller number of uncorrelated variables called principal components.

PRINCIPAL COMPONENT ANALYSIS (PCA) (contd…) The first principal component accounts for as much of the variability in the data as possible, and each succeeding component accounts for as much of the remaining variability as possible. Since images are array of data points with each point representing color, PCA can be used for reducing the image data (extracting features) to smaller dimension to represent the image qualities. remaining variability as possible. The reduced feature represents the spatial distribution of the pixel gray values.

PRINCIPAL COMPONENT ANALYSIS (PCA) (contd…) PCA (Principal Component Analysis): – Project the samples (points) perpendicularly onto the axis of ellipsoid – Rotates the ellipsoid to be parallel to the coordinate axes – Use the fewer and more important coordinates to represent the original samples Transforms of PCA: The first few eigenvectors of the covariance matrix

BACKGROUND LEARNING FOR LETTER-DETECTION Given a document image we first convert it to a gray level image. Since we are working with local texture representation only, this is not going to effect the processing of the image. Then we divide the document image into sub-images where all sub-images are non-overlapping blocks of a specific size (We intend to use height = 32 and width = 32 pixels) Normalize each sub-image independent of the other sub- images by subtracting the mean of the sub-image from each pixel.

BACKGROUND LEARNING FOR LETTER-DETECTION (contd…) Normalizing would help in getting rid of any deviation that a specific sub-image might have from the other sub- images, for instance difference in brightness. We then wish to use the sub-images to compute the principal components using PCA. Use first few principal components to obtain a projection matrix to project each sub-image to an n-dimensional vector that constitutes its texture representation. The number of principal components to be used would be decided on an image-to-image basis.

BACKGROUND LEARNING FOR LETTER-DETECTION (contd…) We now project all sub-images to their texture representation as n-dimensional vectors. Now we use this background learning to exclude background sub-images from further image processing. The remaining sub-images are the informative ones. We can now use the remaining sub-images for letter detection. In order to exclude background sub-images, we use k- means clustering on the n-dimensional vectors corresponding to the sub-images.

BACKGROUND LEARNING FOR LETTER-DETECTION (contd…) We developed MATLAB programs to do the above- mentioned processing and then use it on several document images to compare the performance of this procedure and try to further improve it. The cluster corresponding to the maximum number of sub-images represents the background. By removing these sub-images, we would be reducing the background and thus reduce total scan area for OCR software. We approximate k by observing the resultant image from PCA and vary k by trial and error method.

RESULTS :  Original Document  After PCA with sub-image size 16 and 25 first PCA components

RESULTS…  Sub-image size 16 using all principal components  Sub-image size 32 using all principal components

RESULTS…  k-means with k = 4  k-means with k = 8

RESULTS…  k-means with k = 12  k-means with k = 15

RESULTS… Text Image after removing background

RESULTS…  Original Document  After PCA with sub-image size 32 and 80 first PCA components

RESULTS…  Sub-image size 16 using all principal components  Sub-image size 32 using all principal components

RESULTS…  k-means with k = 6  k-means with k = 12

RESULTS…  k-means with k = 18  k-means with k = 30

RESULTS… Text Image after removing background

RESULTS…  Original Document  After PCA with sub-image size 16

RESULTS…  k-means with k = 6  k-means with k = 30

RESULTS… Text Image after removing background

RESULTS…  Original Document  After PCA with sub-image size 16

RESULTS…  k-means with k = 6  k-means with k = 30

RESULTS… Text Image after removing background

RESULTS…  Original Document  After PCA with sub-image size 16

RESULTS… Text Image after removing background

ANALYSIS & CONCLUSION We tried with sub-images of different sizes: 16, 32, 64 etc. Initially we were under the impression that smaller sub-images would perform better but take more time to execute. We figured out that is not necessarily true and that depends on the image.

ANALYSIS & CONCLUSION.. PCA seemed to be pretty successful in identifying the text blocks in the images. In most of the images we used, we got a pretty good success rate using the sub-images as the training set. The cluster corresponding to the maximum number of sub-images was the background in all cases.

ANALYSIS & CONCLUSION More number of clusters doesn’t necessarily produces more text. Some text which was visible with less number of clusters, wasn’t visible with more. However, more number of clusters reduced removed background. There seems to be a trade off and an optimal cluster size specific to each image.

ANALYSIS & CONCLUSION Removing the background by replacing the cluster corresponding to the maximum number of sub-images seems to be a pretty good method of reducing space to be scanned by OCR. The number of clusters to be used is very much image dependent. The image produced by the PCA gives some idea.

ANALYSIS & CONCLUSION In all our images PCA followed by clustering was successful in removing some background space. It also seem to do a pretty good work of image detection. We conclude that this method works and could be used as a tool to reduce space to be scanned by OCR.

REFERENCES Image Retrieval Using Local PCA Texture Representation by Longin Jan Latecki, Venugopal Rajagopal, Ari Gross Gonzalez, Woods, Eddins. Digital Image Processing Using MATLAB Web material and Notes.