1 Recognition of Multi-Fonts Character in Early-Modern Printed Books Chisato Ishikawa(1), Naomi Ashida(1)*, Yurie Enomoto(1), Masami Takata(1), Tsukasa.

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

UNIVERSIDAD DE MURCIA LÍNEA DE INVESTIGACIÓN DE PERCEPCIÓN ARTIFICIAL Y RECONOCIMIENTO DE PATRONES - GRUPO DE COMPUTACIÓN CIENTÍFICA A CAMERA CALIBRATION.
Michele Merler Jacquilene Jacob.  Applications online are inherently insecure  Growing rate of hackers  Confidentiality of online systems should be.
QR Code Recognition Based On Image Processing
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
Face Recognition and Biometric Systems Eigenfaces (2)
Segmentation of Touching Characters in Devnagari & Bangla Scripts Using Fuzzy MultiFactorial Analysis Presented By: Sanjeev Maharjan St. Xavier’s College.
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
M.S. Student, Hee-Jong Hong
Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers Recognition of Fragmented Characters Using Multiple Feature-Subset Classifiers.
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
Chapter 1: Introduction to Pattern Recognition
Handwritten Character Recognition Using Artificial Neural Networks Shimie Atkins & Daniel Marco Supervisor: Johanan Erez Technion - Israel Institute of.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Highlights Lecture on the image part (10) Automatic Perception 16
California Car License Plate Recognition System ZhengHui Hu Advisor: Dr. Kang.
Handwritten Thai Character Recognition Using Fourier Descriptors and Robust C-Prototype Olarik Surinta Supot Nitsuwat.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Oral Defense by Sunny Tang 15 Aug 2003
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Dan Simon Cleveland State University
Presented By: Dahan Yakir Sepetnitsky Vitali. 2  The will to explore mathematical expressions given as a printed or captured image  It would be nice.
Large-Scale Content-Based Image Retrieval Project Presentation CMPT 880: Large Scale Multimedia Systems and Cloud Computing Under supervision of Dr. Mohamed.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Multiclass object recognition
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
FEATURE EXTRACTION FOR JAVA CHARACTER RECOGNITION Rudy Adipranata, Liliana, Meiliana Indrawijaya, Gregorius Satia Budhi Informatics Department, Petra Christian.
Presented by Tienwei Tsai July, 2005
BACKGROUND LEARNING AND LETTER DETECTION USING TEXTURE WITH PRINCIPAL COMPONENT ANALYSIS (PCA) CIS 601 PROJECT SUMIT BASU FALL 2004.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Ajay Kumar, Member, IEEE, and David Zhang, Senior Member, IEEE.
K. Zagoris, K. Ergina and N. Papamarkos Image Processing and Multimedia Laboratory Department of Electrical & Computer Engineering Democritus University.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Compiled By: Raj G Tiwari.  A pattern is an object, process or event that can be given a name.  A pattern class (or category) is a set of patterns sharing.
Avoiding Segmentation in Multi-digit Numeral String Recognition by Combining Single and Two-digit Classifiers Trained without Negative Examples Dan Ciresan.
COMPARISON OF IMAGE ANALYSIS FOR THAI HANDWRITTEN CHARACTER RECOGNITION Olarik Surinta, chatklaw Jareanpon Department of Management Information System.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
Handwritten Recognition with Neural Network Chatklaw Jareanpon, Olarik Surinta Mahasarakham University.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Mobile Image Processing
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Zhongyan Liang, Sanyuan Zhang Under review for Journal of Zhejiang University Science C (Computers & Electronics) Publisher: Springer A Credible Tilt License.
1 Machine Vision. 2 VISION the most powerful sense.
Scanned Documents INST 734 Module 10 Doug Oard. Agenda Document image retrieval  Representation Retrieval Thanks for David Doermann for most of these.
A Lawn Deterioration Model Constructed from Image Data Yurie Enomoto, Chisato Ishikawa, Masami Takata, Kazuki Joe Department of Advanced Information &
Preliminary Transformations Presented By: -Mona Saudagar Under Guidance of: - Prof. S. V. Jain Multi Oriented Text Recognition In Digital Images.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
1 A Statistical Matching Method in Wavelet Domain for Handwritten Character Recognition Presented by Te-Wei Chiang July, 2005.
An intelligent strategy for checking the annual inspection status of motorcycles based on license plate recognition Yo-Ping Huang a, Chien-Hung Chen b,
Optical Character Recognition
OCR Reading.
S.Rajeswari Head , Scientific Information Resource Division
Introduction Machine Learning 14/02/2017.
Bag-of-Visual-Words Based Feature Extraction
Extracting Old Persian Cuneiform Font Out of
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Final Year Project Presentation --- Magic Paint Face
Brain Hemorrhage Detection and Classification Steps
UN Workshop on Data Capture, Bangkok Session 7 Data Capture
Web Document Image Analysis (OCR) Servers
Deep Learning Hierarchical Representations for Image Steganalysis
Applications of Cellular Neural Networks to Image Understanding
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
UN Workshop on Data Capture, Dar es Salaam Session 7 Data Capture
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Support vector machine-based text detection in digital video
A Novel Smoke Detection Method Using Support Vector Machine
Presentation transcript:

1 Recognition of Multi-Fonts Character in Early-Modern Printed Books Chisato Ishikawa(1), Naomi Ashida(1)*, Yurie Enomoto(1), Masami Takata(1), Tsukasa Kimesawa(2) and Kazuki Joe(1) (1) Nara Women ’ s University, Japan (2) National Diet Library, Japan * Currently work for Mitsubishi Electric co

2 Contents Introduction Multi-fonts character recognition –Feature extraction from character images –Learning method for feature Experiments –Improvement of pre-process Conclusions and future work

3 Introduction The Digital Library from the Meiji Era (Supported by the National Diet Library in Japan) –Digital archive: Books published in the Meiji and Taisho eras Top pageData Viewer Search box The digital data are opened at the project Web site

4 Image data Introduction Main bodies of books Development of an OCR for multi-fonts character in early-modern printed books Our goal –Too many kinds of fonts –Existence of old characters –Very noisy image Existence OCRs are not applicable. Full text search, text function: Not supported Conversion Text data

5 Flow of OCR Input image data Character image Pre-process Character image data X Preprocessed image data X’ Feature extraction Feature vector v Recognition Recognized class no. n Contents of this presentation

6 Flow of our OCR Pre-process Noise reduction Normalization – Removing margin – Normalizing size – Normalizing position Input image data Character image Pre-process Character image data X Preprocessed image data X’ Feature extraction Feature vector v Recognition Recognized class no. n

7 Feature extraction Flow of our OCR Feature Extraction Extraction of a PDC feature Peripheral Direction Contributivity Reflects four statuses of character-lines: ・ Direction ・ Connectivity ・ Relative position ・ Complexion Input image data Character image Preprocessing Preprocessed image data X’ Feature vector v Recognition Recognized class no. n Character image data X

8 PDC Feature Scanning from 8 directions Scanning-line Scanning the lengths of connected black dots for 4 directions Reflecting the direction and the connectivity of character-lines Reflecting the position of character-lines Target character image Direction contributivity is calculated from the scanned lengths A vector of 4 elements

9 are not 0 → Complex character-lines are 0 → Simple character-lines Deeper level’s PDC Feature Scanning-line 1st depth 2nd depth 3rd depth 1st depth 2nd depth3rd depth Black dot: Direction contributivity is not 0 Base image Scanning-line Reflecting the complexity of character-lines Direction contributivity

10 PDC Feature PDC feature vector: Direction contributivities set Resolution: 16 Depth: 3 Direction: 8 Direction contributivity element: 4 Direction(8)*Resolution(16)*Depth(3)*Element(4)=1536 Dimension number=

11 Recognition Flow of our OCR Recognition Recognition by an SVM Support Vector Machine –High generalization capability –Independence of the number of target vector dimension –Low calculation cost Input image data Character image Preprocessing Preprocessed image data X’ Feature extraction Feature vector v Recognized class no. n A character image data X

12 Experiments Experimental sample data –Character images obtained from “The Digital Library from the Meiji era” –Target characters :

13 Examples of Sample Images No.1 ( 行 ) No.2 ( 三 )No.3 ( 人 ) No.4 ( 生 ) No.5 ( 十 )No.6 ( 來 )No.7 ( 小 ) No.8 ( 中 )No.9 ( 年 )No.10 ( 彼 ) Monochrome or 256-grayscale

14 PDC feature Experiments Description(1/2) Conversion of character images to feature vectors –Pre-process 1.Binarization Threshold: Noise Reduction Median filter (Filter size : 3×3) 3.Normalization Removing margin and scaling to 128×128 –Extraction of PDC features Vector dimension: Pre-process PDC feature Extraction of PDC features 1.2.

15 Experiments Description(2/2) Learning and evaluation of a recognition model –Learning recognition model with training samples to SVM Used SVM: LIB-SVM Parameters of SVM: Tweaked by grid search –Evaluation of the recognition model by using test samples PDC feature PDC feature Test samples PDC feature Training samples SVM (LIB-SVM) Learning Tweaked by grid-search Parameters Evaluation Recognition model 50 samples for each character

16 Result of Recognition Model Evaluation Recognition rate: 97.8% cf. Recognition rate by neural network(NN) ・・ 77.6% Computation time ・・ SVM: NN= 1 : 7.7 ※ We have shown this result at 73th Mathematical Modeling and Problem Solving (MPS) in March, 2009.

17 Recognition Error in Result Some images are not recognized because of … noise similarity of character forms Diminishable by an improvement of pre-process or

18 Pre-process 1.Binarization Threshold:t=128 2.First noise reduction Median filter , Filter size : 3×3 3.Normalization 4.Second noise reduction Based on estimated width of character-line 5.Normalization Improvement of Pre-process Discriminant Analysis

19 pipi pjpj lp i lp j Estimation of line width by using the largest connected component X lp n : Length of the shortest connected line pass through pixel p n (p n ⊂ X) Elimination of connected component whose area is smaller than The largest component X Target image Estimated width of character-line: b=median value of lp n Noise Reduction based on Estimation of Character-line Width

20 Target image Estimation of line width by using the largest connected component X lp n : Length of the shortest connected line pass through pixel p n (p n ⊂ X) Elimination of connected components whose area are smaller than Estimated width of character-line: b=median value of lp n Noise Reduction based on Estimation of Character-line Width

21 Noise Reduction based on Estimation of Character-line Width Target image Estimation of line width by using the largest connected component X lp n : Length of the shortest connected line pass through pixel p n (p n ⊂ X) Elimination of connected components whose area are smaller than Estimated width of character-line: b=median value of lp n

22 Result of Improved Pre-process Adoption Recognition rate 97.8%→99.0%

23 Discussion Case: better recognition( Error→Correct) Previous pre-process Error Improved pre-process Correct Quality of test samples are improved Quality of training samples are improved More efficient recognition model

24 Residual noise Error Similar form of character no.5( 十 ) Error Major form of no.8 Shorter than major form →Similar with one horizontal line Previous Error Improved Error Connected to character-line Discussion Case: unchanged( Error→Error)

25 Discussion Case: worse recognition ( Correct→Error) PreviousImproved Training samples with lack of line are reduced Recognition rate of data with lack of line becomes low Previous Correct Improved Error Pre-processed images

26 Conclusions and Future work Recognition of multi-fonts character in Early-Modern Printed Books –Proposal of our method which uses PDC feature and SVM –Experimentations of applying our method The results show high recognition rate Improvement of noise reduction leads higher recognition rate –Recognized 10 kinds of character at 99 % accuracy Future works –Dealing lots of character kinds Recognition of similar form characters –Automation of extracting character area Hierarchical recognition method

27 Thank you for your attention!