Empirical Evaluation of Dissimilarity Measures for Color and Texture

Slides:



Advertisements
Similar presentations
Similarity and Difference
Advertisements

QR Code Recognition Based On Image Processing
Evaluating Color Descriptors for Object and Scene Recognition Koen E.A. van de Sande, Student Member, IEEE, Theo Gevers, Member, IEEE, and Cees G.M. Snoek,
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Robust statistical method for background extraction in image segmentation Doug Keen March 29, 2001.
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Audio Meets Image Retrieval Techniques Dave Kauchak Department of Computer Science University of California, San Diego
Computer Vision Lecture 16: Texture
The Image Histogram.
Visual matching: distance measures
Image Processing David Kauchak cs160 Fall 2009 Empirical Evaluation of Dissimilarity Measures for Color and Texture Jan Puzicha, Joachim M. Buhmann, Yossi.
Image Similarity and the Earth Mover’s Distance Empirical Evaluation of Dissimilarity Measures for Color and Texture Y. Rubner, J. Puzicha, C. Tomasi and.
Discriminative and generative methods for bags of features
Computer Vision Group, University of BonnVision Laboratory, Stanford University Abstract This paper empirically compares nine image dissimilarity measures.
Young Deok Chun, Nam Chul Kim, Member, IEEE, and Ick Hoon Jang, Member, IEEE IEEE TRANSACTIONS ON MULTIMEDIA,OCTOBER 2008.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Texture-Based Image Retrieval for Computerized Tomography Databases Winnie Tsang, Andrew Corboy, Ken Lee, Daniela Raicu and Jacob Furst.
Segmentation Divide the image into segments. Each segment:
Clustering Color/Intensity
Texture Readings: Ch 7: all of it plus Carson paper
Object Recognition Using Geometric Hashing
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Texture-based Deformable Snake Segmentation of the Liver Aaron Mintz Daniela Stan Raicu, PhD Jacob Furst, PhD.
Heather Dunlop : Advanced Perception January 25, 2006
Content-Based Image Retrieval using the EMD algorithm Igal Ioffe George Leifman Supervisor: Doron Shaked Winter-Spring 2000 Technion - Israel Institute.
Image Processing David Kauchak cs458 Fall 2012 Empirical Evaluation of Dissimilarity Measures for Color and Texture Jan Puzicha, Joachim M. Buhmann, Yossi.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Computer Vision James Hays, Brown
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
Nearest Neighbor (NN) Rule & k-Nearest Neighbor (k-NN) Rule Non-parametric : Can be used with arbitrary distributions, No need to assume that the form.
Image Classification 영상분류
Digital Image Processing CCS331 Relationships of Pixel 1.
1 Multiple Classifier Based on Fuzzy C-Means for a Flower Image Retrieval Keita Fukuda, Tetsuya Takiguchi, Yasuo Ariki Graduate School of Engineering,
A Region Based Stereo Matching Algorithm Using Cooperative Optimization Zeng-Fu Wang, Zhi-Gang Zheng University of Science and Technology of China Computer.
CS654: Digital Image Analysis
Computer Graphics and Image Processing (CIS-601).
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Query Sensitive Embeddings Vassilis Athitsos, Marios Hadjieleftheriou, George Kollios, Stan Sclaroff.
Content-Based Image Retrieval (CBIR) By: Victor Makarenkov Michael Marcovich Noam Shemesh.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
CSE 185 Introduction to Computer Vision Feature Matching.
1 Some Guidelines for Good Research Dr Leow Wee Kheng Dept. of Computer Science.
A New Method for Crater Detection Heather Dunlop November 2, 2006.
Similarity Measures Spring 2009 Ben-Gurion University of the Negev.
A Statistical Approach to Texture Classification Nicholas Chan Heather Dunlop Project Dec. 14, 2005.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Similarity Measurement and Detection of Video Sequences Chu-Hong HOI Supervisor: Prof. Michael R. LYU Marker: Prof. Yiu Sang MOON 25 April, 2003 Dept.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
CS598:Visual information Retrieval
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
南台科技大學 資訊工程系 Region partition and feature matching based color recognition of tongue image 指導教授:李育強 報告者 :楊智雁 日期 : 2010/04/19 Pattern Recognition Letters,
Efficient Image Classification on Vertically Decomposed Data
Ch8: Nonparametric Methods
PRESENTED BY Yang Jiao Timo Ahonen, Matti Pietikainen
The Earth Mover's Distance
Efficient Image Classification on Vertically Decomposed Data
K Nearest Neighbor Classification
REMOTE SENSING Multispectral Image Classification
Computer Vision Lecture 16: Texture II
A Fast and Scalable Nearest Neighbor Based Classification
بازیابی تصاویر بر اساس محتوا
Nearest Neighbors CSC 576: Data Mining.
EM Algorithm and its Applications
Presentation transcript:

Empirical Evaluation of Dissimilarity Measures for Color and Texture Jan Puzicha, Joachim M. Buhmann, Yossi Rubner & Carlo Tomasi Presented by: Dave Kauchak Department of Computer Science University of California, San Diego dkauchak@cs.ucsd.edu

The Problem: Image Dissimilarity ) = ? ,

Where does this problem arise in computer vision? Image Classification Image Retrieval Image Segmentation

Classification ? ? ?

Retrieval Jeremy S. De Bonet, Paul Viola (1997). Structure Driven Image Database Retrieval. Neural Information Processing 10 (1997).

Segmentation http://vizlab.rutgers.edu/~comanici/segm_images.html

Histograms for image dissimilarity Examine the distribution of features, rather than the features themselves General purpose (i.e. any distribution of features) Resilient to variations (shadowing, changes in illumination, shading, etc.) Can use previous work in statistics, etc.

Histogram Example

Histogramming Image Features Color Texture Shape Others… Create histogram through binning or some procedure to get a distribution

Color Which is more similar? L*a*b* was designed to be uniform in that perceptual “closeness” corresponds to Euclidean distance in the space.

L*a*b* L – lightness (white to black) a – red-greeness b – yellowness-blueness Nice features: easy to adjust contrast

Texture Texture is not pointwise like color Texture involves a local neighborhood Gabor Filters are commonly used to identify texture features

Gabor Filters Gabor filters are Gaussians modulated by sinusoids They can be tuned in both the scale (size) and the orientation A filter is applied to a region and is characterized by some feature of the energy distribution (often mean and standard deviation)

Examples of Gabor Filters Scale: 3 at 72° Scale: 4 at 108° Scale: 5 at 144°

Creating Histograms from Features Regular Binning Simple Choosing bins important. Bins may be too large or too small Adaptive Binning Bins are adapted to the distribution (usually using some form of K-means)

Marginal Histograms Marginal histograms only deal with a single feature Normal Binning Marginal binning resulting in 2 histograms

Cumulative Histogram Normal Histogram Cumulative Histogram

Dissimilarity Measure Using the Histograms Heuristic Histogram Distances Non-parametric Test Statistics Information-Theoretic diverges Ground distance measures

Notation D(I,J) is the dissimilarity of images I and J f(i;J) is histogram entry i in histogram of image J fr(i;J) is marginal histogram entry i of image J Fr(i;J) is the cumulative histogram

Heuristic Histogram Distances Minkowski-form distance Lp Special cases: L1: absolute, cityblock, or Manhattan distance L2: Euclidian distance L: Maximum value distance

More heuristic distances Weighted-Mean-Variance (WMV) Only includes minimal information about distribution

Non-parametric Test Statistics Kolmogorov-Smirnov distance (K-S) Cramer/von Mises type (CvM)

Cumulative Difference Example Histogram 1 Histogram 2 Difference - = CvM = K-S =

Non-parametric Test Statistics (cont.) 2-statistic (chi-square) Simple statistical measure to decide if two samples came from the same underlying distribution

Information-Theoretic diverges How well can one distribution be coded using the other as a codebook? Kullback-Leibler divergence (KL) Jeffrey-divergence (JD)

Ground Distance Measure Based on some metric of distance between individual features Earth Movers Distance (EMD) Minimal cost to transform one distribution to the other Only measure that works on distributions with a different number of bins Give an general description…

EMD One distribution can be seen as a mass of earth properly spread in space, the other as a collection of holes in that same space Distributions are represented as a set of clusters and an associated weight Computing the dissimilarity then becomes the transportation problem

Transportation Problem Some number of suppliers with goods Some other number of consumers wanting goods Each consumer-supplier pair has an associated cost to deliver one unit of the goods Find least expensive flow of goods from supplier to consumer

Various properties of the metrics K-S, CvM and WMV are only defined for marginal distributions Lp, WMV, K-S, CvM and, under constraints, EMD all obey the triangle inequality WMV is particularly quick because the calculation is quick and the values can be pre-computed offline EMD is the most computationally expensive

Key Components for Good Comparison Meaningful quality measure Subdivision into various tasks/applications (classification, retrieval and segmentation) Wide range of parameters should be measured An uncontroversial “ground truth” should be established

Data Set: Color Randomly chose 94 images from set of 2000 94 images represent separate classes Randomly select disjoint set of pixels from the images Set size of 4, 8, 16, 32, 64 pixels 16 disjoint samples per set per image

Data Set: Texture Brodatz album Collection of wide range of texture (e.g. cork, lawn, straw, pebbles, sand, etc.) Each image is considered a class (as in color) Extract sets of 16 non-overlapping blocks sizes 8x8, 16x16,…, 256x256

Setup: Classification k-Nearest Neighbor classifier is used Nearest Neighbor classification: given a collection of labeled points S and a query point q, what point belonging to S is closest to q? k nearest is a majority vote of the k closest points k = 1, 3, 5 and 7 Average misclassification rate percentage using leave-one-out

Setup: Classification (cont.) Bins: {4, 8, 16, 32, 64, 128, 256} Texture case, three sets of filters were used of sizes 12, 24 and 40 filters 1000 CPU hours of computation

Results: Classification, color data set

Results: Classification, texture data set

Results: Classification For small sample sizes, the WMV measure performs best in the texture case. WMV only estimates means and variances Less sensitive to sampling noise EMD also performs well for small sample sizes Local binning provides additional information For large sample sizes 2 test performs best

Results: Classification (cont.) For texture classification, marginal distributions do better than multidimensional distributions except for very large sample sizes (256x256) Binning is not well adapted to the data since it is fixed for all the 94 classes EMD, which uses local adaption does much better For multidimensional histograms, the more bins the better the performance For texture, usually 12 filters is enough

Setup: Image Retrieval Vary sample size Vary number of images retrieved Performance measured based on precision (i.e. percent correct of the images retrieved) vs. the number of images retrieved

Results: Image Retrieval

Results: Image Retrieval (cont.) Similar to classification EMD, WMV, CvM and K-S performed well for small sample sizes JD, 2 and KL perform better for larger sizes

Setup: Segmentation 100 images Each image consists of 5 different Brodatz textures For multivariate, the bins are adapted to specific image

Setup: Segmentation (cont.) Image is divided into 16384 sites (128 x 128 grid) A histogram is calculate for each site Each site histogram is then compared with 80 randomly selected sites Image sites with high average similarity are then grouped

Results: Segmentation

Results: Segmentation (cont.) Median 20% Quantile L1 marginal 8.2% 12% 2 marginal 8.1% 13% JD marginal KS marginal 10.8% 20% CvM marginal 10.9% 22% L1 full 6.8% 9% 2 full 6.6% 10% JD full

Results: Segmentation (cont.) Binning can be adapted to image Increased accuracy in representing multidimensional distributions Adaptive multivariate outperforms marginal Best results were obtained by 2 EMD suffers from high computational complexity

Conclusions No measure is overall best Marginal histograms and aggregate measures are best for large feature spaces Multivariate histograms perform well with large sample sizes EMD performs generally well for both classification and retrieval, but with a high computational cost 2 is a good overall metric (particularly for larger sample sizes