Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.

Slides:



Advertisements
Similar presentations
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Advertisements

A Novel Approach of Assisting the Visually Impaired to Navigate Path and Avoiding Obstacle-Collisions.
IntroductionIntroduction AbstractAbstract AUTOMATIC LICENSE PLATE LOCATION AND RECOGNITION ALGORITHM FOR COLOR IMAGES Kerem Ozkan, Mustafa C. Demir, Buket.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Foreground Background detection from video Foreground Background detection from video מאת : אבישג אנגרמן.
EVENTS: INRIA Work Review Nov 18 th, Madrid.
Object Inter-Camera Tracking with non- overlapping views: A new dynamic approach Trevor Montcalm Bubaker Boufama.
Adviser : Ming-Yuan Shieh Student ID : M Student : Chung-Chieh Lien VIDEO OBJECT SEGMENTATION AND ITS SALIENT MOTION DETECTION USING ADAPTIVE BACKGROUND.
A Colour Face Image Database for Benchmarking of Automatic Face Detection Algorithms Prag Sharma, Richard B. Reilly UCD DSP Research Group This work is.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Video Object Tracking and Replacement for Post TV Production LYU0303 Final Year Project Spring 2004.
Text Detection in Video Min Cai Background  Video OCR: Text detection, extraction and recognition  Detection Target: Artificial text  Text.
ADVISE: Advanced Digital Video Information Segmentation Engine
Face Detection: a Survey Speaker: Mine-Quan Jing National Chiao Tung University.
Region-Level Motion- Based Background Modeling and Subtraction Using MRFs Shih-Shinh Huang Li-Chen Fu Pei-Yung Hsiao 2007 IEEE.
Quadtrees, Octrees and their Applications in Digital Image Processing
LYU 0102 : XML for Interoperable Digital Video Library Recent years, rapid increase in the usage of multimedia information, Recent years, rapid increase.
Domenico Bloisi, Luca Iocchi, Dorothy Monekosso, Paolo Remagnino
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
Tracking Video Objects in Cluttered Background
Automatic Camera Calibration for Image Sequences of a Football Match Flávio Szenberg (PUC-Rio) Paulo Cezar P. Carvalho (IMPA) Marcelo Gattass (PUC-Rio)
Smart Traveller with Visual Translator. What is Smart Traveller? Mobile Device which is convenience for a traveller to carry Mobile Device which is convenience.
Triangle-based approach to the detection of human face March 2001 PATTERN RECOGNITION Speaker Jing. AIP Lab.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
© 2013 IBM Corporation Efficient Multi-stage Image Classification for Mobile Sensing in Urban Environments Presented by Shashank Mujumdar IBM Research,
Computer Vision Systems for the Blind and Visually Disabled. STATS 19 SEM Talk 3. Alan Yuille. UCLA. Dept. Statistics and Psychology.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
BlindAid Semester Final Presentation Sandra Mau, Nik Melchior, and Maxim Makatchev.
(CONTROLLER-FREE GAMING
Abstract Some Examples The Eye tracker project is a research initiative to enable people, who are suffering from Amyotrophic Lateral Sclerosis (ALS), to.
Interactive Input Methods & Graphical User Input
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
COMPUTER VISION: SOME CLASSICAL PROBLEMS ADWAY MITRA MACHINE LEARNING LABORATORY COMPUTER SCIENCE AND AUTOMATION INDIAN INSTITUTE OF SCIENCE June 24, 2013.
CGMB214: Introduction to Computer Graphics
Submitted by:- Vinay kr. Gupta Computer Sci. & Engg. 4 th year.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Digital Image Processing & Analysis Spring Definitions Image Processing Image Analysis (Image Understanding) Computer Vision Low Level Processes:
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Gili Werner. Motivation Detecting text in a natural scene is an important part of many Computer Vision tasks.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Digital Image Processing & Analysis Fall Outline Sampling and Quantization Image Transforms Discrete Cosine Transforms Image Operations Image Restoration.
MACHINE VISION Machine Vision System Components ENT 273 Ms. HEMA C.R. Lecture 1.
Expectation-Maximization (EM) Case Studies
Inference in generative models of images and video John Winn MSR Cambridge May 2004.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Wonjun Kim and Changick Kim, Member, IEEE
Target Tracking In a Scene By Saurabh Mahajan Supervisor Dr. R. Srivastava B.E. Project.
Scene Text Extraction Using Focus of Mobile Camera Egyul Kim, SeongHun Lee, JinHyung Kim Artificial Intelligence & Pattern Recognition Lab, KAIST, Korea.
Preliminary Transformations Presented By: -Mona Saudagar Under Guidance of: - Prof. S. V. Jain Multi Oriented Text Recognition In Digital Images.
Computer Vision Computer Vision based Hole Filling Chad Hantak COMP December 9, 2003.
Image Text & Audio hacks. Introduction Image Processing is one of the fastest growing technology in the field of computer science. It is a method to convert.
Processing Images and Video for An Impressionist Effect Automatic production of “painterly” animations from video clips. Extending existing algorithms.
License Plate Recognition of A Vehicle using MATLAB
Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.
Automatic License Plate Recognition for Electronic Payment system Chiu Wing Cheung d.
Video object segmentation and its salient motion detection using adaptive background generation Kim, T.K.; Im, J.H.; Paik, J.K.;  Electronics Letters 
Input and output devices for visually impaired users
Gait Recognition Gökhan ŞENGÜL.
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Chapter 2: Input and output devices
Machine Vision Acquisition of image data, followed by the processing and interpretation of these data by computer for some useful application like inspection,
Vehicle Segmentation and Tracking in the Presence of Occlusions
Text Detection in Images and Video
Eric Grimson, Chris Stauffer,
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
AHED Automatic Human Emotion Detection
AHED Automatic Human Emotion Detection
A Novel Smoke Detection Method Using Support Vector Machine
Presentation transcript:

Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons

EXISTING METHOD Walking safely and confidently without any human assistnce in urban or unknown environments is a difficult task for blind people. Visually impaired people generally use either the typical white cane or the guide dog to travel independently. But these methods are used only to guide blind people for safe path movement. but these cannot provide any product assistance like shopping………..

PROPOSED METHOD We propose a camera-based label reader to help blind persons to read names of labels on the products. Camera acts as main vision in detecting the label image of the product or board then image is processed internally . And separates label from image , and finally identifies the product and identified product name is pronounced through voice. Then received label image is converted to text . Once the identified label name is converted to text and converted text is displayed on display unit connected to controller. Now converted text should be converted to voice to hear label name as voice through ear phones connected to audio.

FRAMEWORK AND ALGORITHM OVERVIEW The system framework consists of three functional components: Scene capture Data processing Audio output The scene capture component collects scenes containing objects of interest in the form of images or video. In our prototype, it corresponds to a camera attached to a pair of sunglasses.

CONT…… The data processing component is used for deploying our proposed algorithms, including object-of-interest detection to selectively extract the image of the object held by the blind user from the cluttered background or other neutral objects in the camera view Text localization to obtain image regions containing text, and text recognition to transform image-based text information into readable codes The audio output component is to inform the blind user of recognized text codes. A Bluetooth earpiece with minimicrophone is employed for speech output.

Our main contributions embodied in this prototype system are….. A novel motion-based algorithm to solve the aiming problem for blind users by their simply shaking the object of interest for a brief period. A novel algorithm of automatic text localization to extract text regions from complex background and multiple text patterns; A portable camera-based assistive framework to aid blind persons reading text from hand-held objects.

OBJECT REGION DETECTION To ensure that the hand-held object appears in the camera view, we employ a camera with a reasonably wide angle in our prototype system (since the blind user may not aim accurately) This may result in some other extraneous but perhaps text-like objects appearing in the camera view. To extract the hand-held object of interest from other objects in the camera view, we ask users to shake the hand-held objects containing the text they wish identify. then employ a motion-based method to localize the objects from cluttered background.

Background subtraction (BGS) is a conventional and effective approach to detect moving objects for video surveillance systems with stationary cameras. This method is done based on the frame variations. Since background imagery is nearly constant in all frames, a Gaussian always compatible Its subsequent frame pixel distribution is more likely to be the background model To detect moving objects in a dynamic scene, many adaptive BGS techniques have been developed.

TEXT RECOGNITION AND AUDIO OUTPUT Text recognition is performed by off-the-shelf OCR prior to output of informative words from the localized text regions. A text region labels the minimum rectangular area for the accommodation of characters inside it, So the border of the text region contacts the edge boundary of the text character. OCR generates better performance if text regions are first assigned proper margin areas and binarized to segment text characters from background. Thus, each localized text region is enlarged by enhancing the height and width by pixels, respectively,

ADVANTAGES It removes physical hardware requirement . Portability. Accuracy and Flexibility. Automatic detection

CONCLUSION To read printed text on hand-held objects for assisting blind person In order to solve the common aiming problem for blind users. This method can effectively distinguish the object of interest from background or other objects in the camera view. To extract text regions from complex backgrounds, we have proposed a novel text localization algorithm based on models of stroke orientation and edge distributions. OCR is used to perform word recognition on the localized text regions and transform into audio output for blind users.

FUTURE WORK Our future work will extend our localization algorithm to process text strings with characters fewer than three and to design more robust block patterns for text feature extraction. We will also extend our algorithm to handle non horizontal text strings. Furthermore, we will address the significant human interface issues associated with reading text by blind users.

REFERENCES World Health Organization. (2013). 10 facts about blindness and visual impairment [Online]. Available: www.who.int/features/factfiles/blindness/blindness_f acts/en/index.html[28] C. Stauffer and W. E. L. Grimson, “Adaptive background mixture models for real-time tracking,” presented at the IEEE Comput. Soc.Conf.Comput. Vision Pattern Recognit., Fort Collins, CO, USA, 2013