Presented By: Vennela Sunnam

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Ter Haar Romeny, ICPR 2010 Introduction to Scale-Space and Deep Structure.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
Image Repairing: Robust Image Synthesis by Adaptive ND Tensor Voting IEEE Computer Society Conference on Computer Vision and Pattern Recognition Jiaya.
Presented by Xinyu Chang
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Computer Vision Lecture 16: Texture
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
Complex Feature Recognition: A Bayesian Approach for Learning to Recognize Objects by Paul A. Viola Presented By: Emrah Ceyhan Divin Proothi Sherwin Shaidee.
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
3-D Depth Reconstruction from a Single Still Image 何開暘
1 Texture Texture is a description of the spatial arrangement of color or intensities in an image or a selected region of an image. Structural approach:
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Texture Texture is a description of the spatial arrangement of color or intensities in an image or a selected region of an image. Structural approach:
Segmentation Divide the image into segments. Each segment:
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Visual Information Extraction in Content-based Image Retrieval System Presented by: Mian Huang Weichuan Dong Apr 29, 2004.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Texture Readings: Ch 7: all of it plus Carson paper
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
A fuzzy video content representation for video summarization and content-based retrieval Anastasios D. Doulamis, Nikolaos D. Doulamis, Stefanos D. Kollias.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Copyright © 2012 Elsevier Inc. All rights reserved.
Computer vision.
Multimodal Interaction Dr. Mike Spann
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Mean-shift and its application for object tracking
October 14, 2014Computer Vision Lecture 11: Image Segmentation I 1Contours How should we represent contours? A good contour representation should meet.
Image Retrieval Part I (Introduction). 2 Image Understanding Functions Image indexing similarity matching image retrieval (content-based method)
Content-Based Image Retrieval
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
EECS 274 Computer Vision Segmentation by Clustering II.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Non-Photorealistic Rendering and Content- Based Image Retrieval Yuan-Hao Lai Pacific Graphics (2003)
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Computer Vision Lecture 6. Probabilistic Methods in Segmentation.
Image Emotional Semantic Query Based On Color Semantic Description Wei-Ning Wang, Ying-Lin Yu Department of Electronic and Information Engineering, South.
Autonomous Robots Vision © Manfred Huber 2014.
CSE 185 Introduction to Computer Vision Feature Matching.
MIT AI Lab / LIDS Laboatory for Information and Decision Systems & Artificial Intelligence Laboratory Massachusetts Institute of Technology A Unified Multiresolution.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Digital Image Processing
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Introduction to Scale Space and Deep Structure. Importance of Scale Painting by Dali Objects exist at certain ranges of scale. It is not known a priory.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
1. 2 What is Digital Image Processing? The term image refers to a two-dimensional light intensity function f(x,y), where x and y denote spatial(plane)
Image Representation and Description – Representation Schemes
SIFT Scale-Invariant Feature Transform David Lowe
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Dynamical Statistical Shape Priors for Level Set Based Tracking
Fitting Curve Models to Edges
Computer Vision Lecture 16: Texture II
Matching Words with Pictures
Blobworld Texture Features
Fourier Transform of Boundaries
EM Algorithm and its Applications
Presentation transcript:

Presented By: Vennela Sunnam 800802277 Blobworld: Image Segmentation Using Expectation-Maximization and Its Application to Image Querying Presented By: Vennela Sunnam 800802277

Table Of Contents Challenge Limitations of Image Retrieval Systems Introduction What is Blobworld? Image Segmentation EM- Algorithm Stages of Blobworld Processing Feature Extraction Extracting Color Features Extracting Texture Features Scale Selection Polarity Conclusion

Challenge Retrieving images from large and varied collections using image content as a key. The image collections are diverse and often poorly indexed; unfortunately, image retrieval systems have not kept pace with the collections they are searching. Approach: Transformation from the raw pixel data to a small set of image regions that are coherent in color and texture.

Limitations of the Image Retrieval Systems Find images containing particular objects based only on their low-level features, with little regard for the spatial organization of those features. Systems based on user querying are often unintuitive.

Introduction Clustering pixels in a joint color-texture-position feature space. Segmentation algorithm is fully automatic and has been run on a collection of 10,000 natural images. The user is allowed to view the internal representation of the submitted image and the query results.

What is BLOBWORLD? A new framework for image retrieval based on segmentation into regions and querying using properties of these regions. The regions generally correspond to objects or parts of objects. Blobworld does not exist completely in the “thing” domain, it recognizes the nature of images as combinations of objects, and querying in Blobworld is more meaningful than it is with simple “stuff” representations.

Image Segmentation Segmentation algorithms make mistakes, causing degradation in performance of any system that uses the segmentation results. As a result, designers of image retrieval systems have generally chosen to use global image properties, which do not depend on accurate segmentation. However, segmenting an image allows us to access the image at the level of objects.

Related Work Color Histograms - encodes the spatial correlation of color-bin pairs Multiresolution wavelet decompositions to perform queries based on iconic matching. EM-Algorithm - estimate the parameters of a mixture of Gaussians model of the joint distribution of pixel color and texture features.

EM- Algorithm In order to segment each image automatically, we model the joint distribution of color, texture, and position features with a mixture of Gaussians. We use the Expectation-Maximization (EM) algorithm to estimate the parameters of this model; the resulting pixel-cluster memberships provide a segmentation of the image. After the image is segmented into regions, a description of each region's color and texture characteristics is produced. In a querying task, the user can access the regions directly, in order to see the segmentation of the query image and specify which aspects of the image are important to the query. When query results are returned, the user also sees the Blobworld representation of each retrieved image; this information assists greatly in refining the query.

Stages of Blobworld Processing From pixels to region descriptions

Feature Extraction Select an appropriate scale for each pixel and extract color, texture, and position features for that pixel at the selected scale. Group pixels into regions by modeling the distribution of pixel features with a mixture of Gaussians using Expectation-Maximization. Describe the color distribution and texture of each region for use in a query.

Extracting Color Features Each image pixel has a three-dimensional color descriptor in the L*a*b* color space. This color space is approximately perceptually uniform; thus, distances in this space are meaningful. We smooth the color features in order to avoid over segmenting regions such as tiger stripes based on local color variation; otherwise, each stripe would become its own region.

Extracting Texture Features Color is a point property, texture is a local neighborhood property. The first requirement could be met to an arbitrary degree of satisfaction by using multi-orientation filter banks such as steerable filters; we chose a simpler method that is sufficient for our purposes. The second requirement, the problem of scale selection, has not received the same level of attention.

Scale Selection Use of a local image property known as polarity. The polarity is a measure of the extent to which the gradient vectors in a certain neighborhood all point in the same direction. The polarity at a given pixel is computed with respect to the dominant orientation in the neighborhood of that pixel.

Fig. 3. Five sample patches from a zebra image. Both (a). 1:5 and (b) Fig. 3. Five sample patches from a zebra image. Both (a) . 1:5 and (b) . 2:5 have stripes (1D flow) of different scales and orientations,(c) is a region of 2D texture with . 1:5, (d) contains an edge with . 0, and (e) is a uniform region with . 0.

Polarity Polarity is defined as:

Factors affecting Polarity Edge: The presence of an edge is signaled by p holding values close to 1 for all . Texture: In regions with 2D texture or 1D flow, p decays with : as the window size increases, pixels with gradients in multiple directions are included in the window, so the dominance of any one orientation decreases. Uniform: When a neighborhood possesses a constant intensity, p takes on arbitrary values since the gradient vectors have negligible magnitudes and arbitrary angles.

Texture Features

Combining Color, Texture, and Position Features The final color/texture descriptor for a given pixel consists of six values: three for color and three for texture. The three color components are the L*a*b* coordinates found after spatial averaging using a Gaussian at the selected scale. The three texture components are ac, pc, and c, computed at the selected scale; the anisotropy and polarity are each modulated by the contrast since they are meaningless in regions of low contrast.

EM Algorithm The EM algorithm is used for finding maximum likelihood parameter estimates when there is missing or incomplete data. The missing data is the Gaussian cluster to which the points in the feature space belong. We estimate values to fill in for the incomplete data (the “E Step”), compute the maximum-likelihood parameter estimates using this data (the “M Step”), and repeat until a suitable stopping criterion is reached. In the case where EM is applied to learning the parameters for a mixture of Gaussians, it turns out that both steps can be combined into a single update step.

Grouping Pixels into Regions

Phases of Grouping Pixels Model Selection Postprocessing Segmentation Results

Model Selection To choose K, the number of mixture components, apply Minimum Description Length(MDL) principle. Choose K to maximize

Postprocessing Perform spatial grouping of those pixels belonging to the same color/texture cluster. We first produce a K-level image which encodes pixel-cluster. Find the color histogram of each region (minus its boundary) using the original pixel colors (before smoothing). For each pixel (in color bin i) on the boundary between two or more regions, reassign it to the region whose histogram value i is largest.

Segmentation Results Large background areas may be arbitrarily split into two regions due to the use of position in the feature vector. The region boundaries sometimes do not follow object boundaries exactly, even when the object boundary is visually quite apparent. This occurs because the color feature is averaged across object boundaries. The object of interest is missed, split, or merged with other regions because it is not visually distinct. In rare cases, a visually distinct object is simply missed. This error occurs mainly when no initial mean falls near the object's feature vectors.

Describing the Regions

Image Retrieval by Querying Two major shortcomings of interfaces are lack of user control and the absence of information about the computer's view of the image.

Querying in Blobworld

Results Distinctive objects Distinctive scenes Distinctive objects and scenes

Content-based Image Retrieval Group pixels into regions which are coherent in low level properties and which generally correspond to objects or parts of objects. Describe these regions in ways that are meaningful to the user. Access these region descriptions, either automatically or with user intervention, to retrieve desired images.

Conclusion Our belief is that segmentation, while imperfect, is an essential first step, as the combinatorics of searching for all possible instances of a class is intractable. A combine architecture for segmentation and recognition is needed, analogous to inference using Hidden Markov Models in speech recognition. We cannot claim that our framework provides an ultimate solution to this central problem in computer vision.

Any Questions?

THANK YOU