In Defense of Nearest-Neighbor Based Image Classification Oren Boiman The Weizmann Institute of Science Rehovot, ISRAEL Eli Shechtman Adobe Systems Inc.

Slides:



Advertisements
Similar presentations
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
Advertisements

Aggregating local image descriptors into compact codes
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
VisualRank: Applying PageRank to Large-Scale Image Search Yushi Jing, Member, IEEE, and Shumeet Baluja, Member, IEEE.
Multi-layer Orthogonal Codebook for Image Classification Presented by Xia Li.
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Ziming Zhang *, Ze-Nian Li, Mark Drew School of Computing Science, Simon Fraser University, Vancouver, B.C., Canada {zza27, li, Learning.
1 Building a Dictionary of Image Fragments Zicheng Liao Ali Farhadi Yang Wang Ian Endres David Forsyth Department of Computer Science, University of Illinois.
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Discriminative and generative methods for bags of features
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Machine Vision and Dig. Image Analysis 1 Prof. Heikki Kälviäinen C50A6100 Lectures 12: Object Recognition Professor Heikki Kälviäinen Machine Vision and.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter 4 (part 2): Non-Parametric Classification
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Multiple Object Class Detection with a Generative Model K. Mikolajczyk, B. Leibe and B. Schiele Carolina Galleguillos.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Class Meeting #11 Data Analysis. Types of Statistics Descriptive Statistics used to describe things, frequently groups of people.  Central Tendency 
Mutual Information-based Stereo Matching Combined with SIFT Descriptor in Log-chromaticity Color Space Yong Seok Heo, Kyoung Mu Lee, and Sang Uk Lee.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 24 – Classifiers 1.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
Svetlana Lazebnik, Cordelia Schmid, Jean Ponce
Yao, B., and Fei-fei, L. IEEE Transactions on PAMI(2012)
Image Classification 영상분류
SVM-KNN Discriminative Nearest Neighbor Classification for Visual Category Recognition Hao Zhang, Alex Berg, Michael Maire, Jitendra Malik.
80 million tiny images: a large dataset for non-parametric object and scene recognition CS 4763 Multimedia Systems Spring 2008.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Classification 1: generative and non-parameteric methods Jakob Verbeek January 7, 2011 Course website:
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Methods for classification and image representation
Linear Models for Classification
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
KNN & Naïve Bayes Hongning Wang Today’s lecture Instance-based classifiers – k nearest neighbors – Non-parametric learning algorithm Model-based.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
CS654: Digital Image Analysis
Object Recognition as Ranking Holistic Figure-Ground Hypotheses Fuxin Li and Joao Carreira and Cristian Sminchisescu 1.
Wonjun Kim and Changick Kim, Member, IEEE
Goggle Gist on the Google Phone A Content-based image retrieval system for the Google phone Manu Viswanathan Chin-Kai Chang Ji Hyun Moon.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
Fast Human Detection in Crowded Scenes by Contour Integration and Local Shape Estimation Csaba Beleznai, Horst Bischof Computer Vision and Pattern Recognition,
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
KNN & Naïve Bayes Hongning Wang
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Learning Mid-Level Features For Recognition
Nonparametric Semantic Segmentation
Video Google: Text Retrieval Approach to Object Matching in Videos
Paper Presentation: Shape and Matching
Digit Recognition using SVMS
Object detection as supervised classification
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
CS 1674: Intro to Computer Vision Scene Recognition
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CS5112: Algorithms and Data Structures for Applications
Video Google: Text Retrieval Approach to Object Matching in Videos
A Graph-Matching Kernel for Object Categorization
Presentation transcript:

In Defense of Nearest-Neighbor Based Image Classification Oren Boiman The Weizmann Institute of Science Rehovot, ISRAEL Eli Shechtman Adobe Systems Inc. & University of Washington Michal Irani The Weizmann Institute of Science Rehovot, ISRAEL

Outline Introduction What degrades NN-image classification? The Approximation Algorithm Using NN Results and Experiments

Introduction The problem of image classification has drawn considerable attention in the Computer Vision community. Image classification methods can be roughly divided into two broad families of approaches (i) Learning-based classifiers (ii) Non-parametric classifiers

Learning-based classifiers Require an intensive learning/training phase of the classifier parameters The leading image classifiers, to date, are learning- based classifiers, in particular SVM-based methods (e.g. parameters of SVM, Boosting, parametric generative models, decision trees, fragments and object parts, etc.)

Non-parametric classifiers That base their classification decision directly on the data, and require no learning/training of parameters. The most common non-parametric method is Nearest- Neighbor-Image Although this is the most popular among the NN-based image classifiers, it provides inferior performance relative to learning-based methods.

Non-parametric classifiers have several very important advantages that are not shared by most learning-based approaches (i) Can naturally handle a huge number of classes. (ii) Avoid overfitting of parameters, which is a central issue in learning based approaches. (iii) Require no learning/training phase.

But the large performance gap between non-parametric NN-image classifiers and learning-based approaches led to non-parametric image classification is not useful. We argue that two practices, that are commonly used in image classification, lead to significant degradation in the performance of non-parametric image classifiers: (i) Descriptor quantization (ii) Computation of ‘Image-to-Image’ distance

Quantization damages non-parametric classifiers Descriptor quantization is often used to generate codebooks (or “bags-of-words”) for obtaining compact image representations Lazebnik et al. [16] further proposed to add rough quantized location information to the histogram representation. The amount of discriminative information is considerably reduced due to the rough quantization. [16] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In IEEE Conference on Computer Vision and Pattern Recognition, 2006.

Highly frequent descriptors have low quantization error, while rare descriptors have high quantization error. The most frequent descriptors in a large database of images (e.g., Caltech-101) comprise of simple edges and corners that appear abundantly in all the classes within the database, and therefore are least informative for classification. In contrast, the most informative descriptors for classification are the ones found in one (or few) class, but are rare in other classes.

Figure 2. Effects of descriptor quantization – Severe drop in descriptor discriminative power. We generated a scatter plot of descriptor discriminative power before and after quantization (for a very large sample set of SIFT descriptors d in Caltech-101, each for its respective class C). We then averaged this scatter plot along the y-axis. This yields the “Average discriminative power after quantization” (the RED graph). The display is in logarithmic scale in both axes. NOTE: The more informative (discriminative) a descriptor d is, the larger the drop in its discriminative power.

Image-to-Image vs. Image-to-Class NN-image classifiers provide good image classification when the query image is similar to one of the labelled images in its class. When there are only few labelled images for classes with large variability in object shape and appearance bad classification is obtained. “Image-to-image” distance becomes the ‘distance’ between two descriptor distributions of the two images (which can be measured via histogram intersection, Chisquare, or KL divergence).

Figure 3. “Image-to-Image” vs. “Image-to-Class” distance. A Ballet class with large variability and small number (three) of ‘labelled’ images (bottom row). Even though the “Query-to-Image” distance is large to each individual ‘labelled’ image, the “Queryto- Class” distance is small. Top right image: For each descriptor at each point in Q we show (in color) the ‘labelled’ image which gave it the highest descriptor likelihood. It is evident that the new query configuration is more likely given the three images, than each individual image seperately.

Probabilistic Formulation

Naive-Bayes classifier  Minimum “Image-to-Class” KL-Distance:

The Approximation Algorithm Using NN Non-Parametric Descriptor Density Estimation The optimal MAP Naive-Bayes image classifier of Eq. (1) requires computing the probability density p(d|C) of descriptor d in a class C. A Parzen density estimation provides an accurate non- parametric approximation of the continuous descriptor probability density p(d|C) [7]. [7] R. Duda, P. Hart, and D. Stork. Pattern Classification. Wiley, New York, 2001.

But this is computationally time-consuming

The NBNN Algorithm

Figure 4. NN descriptor estimation preserves descriptor density distribution and discriminativity. (a) A scatter plot of the 1-NN probability density distribution pNN(d|C) vs. the true distribution p(d|C). Brightness corresponds to the concentration of points in the scatter plot. The plot shows that 1-NN distribution provides a very accurate approximation of the true distribution. (b) 20-NN descriptor approximation (Green graph) and 1- NN descriptor approximation (Blue graph) preserve quite well the discriminative power of descriptors. In contrast, descriptor quantization (Red graph) severely reduces discriminative power of descriptors. Displays are in logarithmic scale in all axes.

Results and Experiments