1 Image Retrieval Hao Jiang Computer Science Department 2009.

Slides:



Advertisements
Similar presentations
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Advertisements

Part 1: Bag-of-words models by Li Fei-Fei (Princeton)
FATIH CAKIR MELIHCAN TURK F. SUKRU TORUN AHMET CAGRI SIMSEK Content-Based Image Retrieval using the Bag-of-Words Concept.
Outline SIFT Background SIFT Extraction Application in Content Based Image Search Conclusion.
Marco Cristani Teorie e Tecniche del Riconoscimento1 Teoria e Tecniche del Riconoscimento Estrazione delle feature: Bag of words Facoltà di Scienze MM.
TP14 - Indexing local features
Large-Scale Image Retrieval From Your Sketches Daniel Brooks 1,Loren Lin 2,Yijuan Lu 1 1 Department of Computer Science, Texas State University, TX, USA.
Multi-layer Orthogonal Codebook for Image Classification Presented by Xia Li.
1 Part 1: Classical Image Classification Methods Kai Yu Dept. of Media Analytics NEC Laboratories America Andrew Ng Computer Science Dept. Stanford University.
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Global spatial layout: spatial pyramid matching Spatial weighting the features Beyond bags of features: Adding spatial information.
Object Recognition. So what does object recognition involve?
Bag-of-features models Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
CVPR 2008 James Philbin Ondˇrej Chum Michael Isard Josef Sivic
Fitting: The Hough transform
Bag of Features Approach: recent work, using geometric information.
Effective Image Database Search via Dimensionality Reduction Anders Bjorholm Dahl and Henrik Aanæs IEEE Computer Society Conference on Computer Vision.
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
Robust and large-scale alignment Image from
Beyond bags of features: Adding spatial information Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
What is Texture? Texture depicts spatially repeating patterns Many natural phenomena are textures radishesrocksyogurt.
Object retrieval with large vocabularies and fast spatial matching
1 Image Recognition - I. Global appearance patterns Slides by K. Grauman, B. Leibe.
Lecture 28: Bag-of-words models
Agenda Introduction Bag-of-words model Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
“Bag of Words”: when is object recognition, just texture recognition? : Advanced Machine Perception A. Efros, CMU, Spring 2009 Adopted from Fei-Fei.
CS335 Principles of Multimedia Systems Content Based Media Retrieval Hao Jiang Computer Science Department Boston College Dec. 4, 2007.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman ICCV 2003 Presented by: Indriyati Atmosukarto.
Beyond bags of features: Adding spatial information Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba.
Object Recognition. So what does object recognition involve?
Bag-of-features models
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
“Bag of Words”: recognition using texture : Advanced Machine Perception A. Efros, CMU, Spring 2006 Adopted from Fei-Fei Li, with some slides from.
Large Scale Recognition and Retrieval. What does the world look like? High level image statistics Object Recognition for large-scale search Focus on scaling.
Keypoint-based Recognition and Object Search
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
Object Recognition and Augmented Reality
Review: Intro to recognition Recognition tasks Machine learning approach: training, testing, generalization Example classifiers Nearest neighbor Linear.
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
CS 766: Computer Vision Computer Sciences Department, University of Wisconsin-Madison Indexing and Retrieval James Hill, Ozcan Ilikhan, Mark Lenz {jshill4,
Indexing Techniques Mei-Chen Yeh.
Clustering with Application to Fast Object Search
Keypoint-based Recognition Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/04/10.
CSE 473/573 Computer Vision and Image Processing (CVIP)
Special Topic on Image Retrieval
Bag-of-features models. Origin 1: Texture recognition Texture is characterized by the repetition of basic elements or textons For stochastic textures,
Svetlana Lazebnik, Cordelia Schmid, Jean Ponce
Video Google: A Text Retrieval Approach to Object Matching in Videos Josef Sivic and Andrew Zisserman.
10/31/13 Object Recognition and Augmented Reality Computational Photography Derek Hoiem, University of Illinois Dali, Swans Reflecting Elephants.
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
Advanced Multimedia Image Content Analysis Tamara Berg.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Lecture 08 27/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CS654: Digital Image Analysis
776 Computer Vision Jan-Michael Frahm Spring 2012.
Video Google: Text Retrieval Approach to Object Matching in Videos Authors: Josef Sivic and Andrew Zisserman University of Oxford ICCV 2003.
Lecture IX: Object Recognition (2)
Highlighted Project 2 Implementations
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CS 2770: Computer Vision Feature Matching and Indexing
Large-scale Instance Retrieval
Video Google: Text Retrieval Approach to Object Matching in Videos
Paper Presentation: Shape and Matching
Large-scale Instance Retrieval
By Suren Manvelyan, Crocodile (nile crocodile?) By Suren Manvelyan,
CS 1674: Intro to Computer Vision Scene Recognition
Video Google: Text Retrieval Approach to Object Matching in Videos
Part 1: Bag-of-words models
Prof. Adriana Kovashka University of Pittsburgh September 17, 2019
Presentation transcript:

1 Image Retrieval Hao Jiang Computer Science Department 2009

Color Histogram Methods Color only schemes tend to find many unrelated images.

Improve Color Histogram Methods  If we can separate the foreground with background the result will be improved. Foreground Background

Improve Color Histogram Methods  Their spatial relations also help to find the right object. Color Blob 2 Color Blob 1

Finding Shapes  Finding similar shapes is a very useful tool in managing large number of images.  Chamfer matching is a standard method to compare the similarity of shapes.  General Hough Transform can also be used to find shapes in images.

Shape Context  Shape context is another widely used feature in shape retrieval. C ij is the distance of shape contexts h i and h j

Improve Matching Efficiency  Fast pruning in matching  Reprehensive shape contexts  Shapemes Greg Mori, Serge Belongie, and Jitendra Malik, Shape Contexts Enable Efficient Retrieval of Similar Shapes, CVPR, 2001

Example Results Reprehensive shape contexts in shape matching

Bag of Words

Slide from Andrew Zisserman Sivic & Zisserman, ICCV 2003

Analogy to documents Of all the sensory impressions proceeding to the brain, the visual experiences are the dominant ones. Our perception of the world around us is based essentially on the messages that reach the brain from our eyes. For a long time it was thought that the retinal image was transmitted point by point to visual centers in the brain; the cerebral cortex was a movie screen, so to speak, upon which the image in the eye was projected. Through the discoveries of Hubel and Wiesel we now know that behind the origin of the visual perception in the brain there is a considerably more complicated course of events. By following the visual impulses along their path to the various cell layers of the optical cortex, Hubel and Wiesel have been able to demonstrate that the message about the image falling on the retina undergoes a step-wise analysis in a system of nerve cells stored in columns. In this system each cell has its specific function and is responsible for a specific detail in the pattern of the retinal image. sensory, brain, visual, perception, retinal, cerebral cortex, eye, cell, optical nerve, image Hubel, Wiesel China is forecasting a trade surplus of $90bn (£51bn) to $100bn this year, a threefold increase on 2004's $32bn. The Commerce Ministry said the surplus would be created by a predicted 30% jump in exports to $750bn, compared with a 18% rise in imports to $660bn. The figures are likely to further annoy the US, which has long argued that China's exports are unfairly helped by a deliberately undervalued yuan. Beijing agrees the surplus is too high, but says the yuan is only one factor. Bank of China governor Zhou Xiaochuan said the country also needed to do more to boost domestic demand so more goods stayed within the country. China increased the value of the yuan against the dollar by 2.1% in July and permitted it to trade within a narrow band, but the US wants the yuan to be allowed to trade freely. However, Beijing has made it clear that it will take its time and tread carefully before allowing the yuan to rise further in value. China, trade, surplus, commerce, exports, imports, US, yuan, bank, domestic, foreign, increase, trade, value ICCV 2005 short course, L. Fei-Fei

Visual words: main idea  Extract some local features from a number of images … e.g., SIFT descriptor space: each point is 128- dimensional Slide credit: D. Nister

Visual words: main idea Slide credit: D. Nister

Visual words: main idea Slide credit: D. Nister

Visual words: main idea Slide credit: D. Nister

Each point is a local descriptor, e.g. SIFT vector.

Slide credit: D. Nister

Visual words  Example: each group of patches belongs to the same visual word Figure from Sivic & Zisserman, ICCV 2003

Visual words 19 Source credit: K. Grauman, B. Leibe More recently used for describing scenes and objects for the sake of indexing or classification. Sivic & Zisserman 2003; Csurka, Bray, Dance, & Fan 2004; many others.

Object Bag of ‘words’ ICCV 2005 short course, L. Fei-Fei

Bags of visual words  Summarize entire image based on its distribution (histogram) of word occurrences.  Analogous to bag of words representation commonly used for documents. 22 Image credit: Fei-Fei Li

Similarly, bags of textons for texture representation Universal texton dictionary histogram Julesz, 1981; Cula & Dana, 2001; Leung & Malik 2001; Mori, Belongie & Malik, 2001; Schmid 2001; Varma & Zisserman, 2002, 2003; Lazebnik, Schmid & Ponce, 2003 Source: Lana Lazebnik

Comparing bags of words  Rank frames by normalized scalar product between their (possibly weighted) occurrence counts---nearest neighbor search for similar images. [ ][ ]

Inverted file index for images comprised of visual words Image credit: A. Zisserman Word number List of image numbers When will this give us a significant gain in efficiency?

 For text documents, an efficient way to find all pages on which a word occurs is to use an index…  We want to find all images in which a feature occurs.  We need to index each feature by the image it appears and also we keep the # of occurrence. Source credit : K. Grauman, B. Leibe Indexing local features: inverted file index

tf-idf weighting  Term frequency – inverse document frequency  Describe frame by frequency of each word within it, downweight words that appear often in the database  (Standard weighting for text retrieval) Total number of words in database Number of occurrences of word i in whole database Number of occurrences of word i in document d Number of words in document d

Slide from Andrew Zisserman Sivic & Zisserman, ICCV 2003 What if query of interest is a portion of a frame? Bags of words for content-based image retrieval

Video Google System 1.Collect all words within query region 2.Inverted file index to find relevant frames 3.Compare word counts 4.Spatial verification Sivic & Zisserman, ICCV 2003  Demo online at : arch/vgoogle/index.html 29 Query region Retrieved frames

30  Collecting words within a query region Query region: pull out only the SIFT descriptors whose positions are within the polygon

Bag of words representation: spatial info  A bag of words is an orderless representation: throwing out spatial relationships between features  Middle ground:  Visual “phrases” : frequently co-occurring words  Semi-local features : describe configuration, neighborhood  Let position be part of each feature  Count bags of words only within sub-grids of an image  After matching, verify spatial consistency (e.g., look at neighbors – are they the same too?)

Visual vocabulary formation Issues:  Sampling strategy: where to extract features?  Clustering / quantization algorithm  Unsupervised vs. supervised  What corpus provides features (universal vocabulary?)  Vocabulary size, number of words 35 K. Grauman, B. Leibe

Sampling strategies 36 Image credits: F-F. Li, E. Nowak, J. Sivic Dense, uniformly Sparse, at interest points Randomly Multiple interest operators To find specific, textured objects, sparse sampling from interest points often more reliable. Multiple complementary interest operators offer more image coverage. For object categorization, dense sampling offers better coverage. [See Nowak, Jurie & Triggs, ECCV 2006]

37 Example: Recognition with Vocabulary Tree  Tree construction: Slide credit: David Nister [Nister & Stewenius, CVPR’06]

38 Vocabulary Tree  Training: Filling the tree Slide credit: David Nister [Nister & Stewenius, CVPR’06]

39 Vocabulary Tree  Training: Filling the tree Slide credit: David Nister [Nister & Stewenius, CVPR’06]

40 Vocabulary Tree  Training: Filling the tree Slide credit: David Nister [Nister & Stewenius, CVPR’06]

41 Vocabulary Tree  Training: Filling the tree Slide credit: David Nister [Nister & Stewenius, CVPR’06]

42 Vocabulary Tree  Training: Filling the tree Slide credit: David Nister [Nister & Stewenius, CVPR’06]

43 Vocabulary Tree  Recognition Slide credit: David Nister [Nister & Stewenius, CVPR’06] RANSAC verification

44 Vocabulary Tree: Performance  Evaluated on large databases  Indexing with up to 1M images  Online recognition for database of 50,000 CD covers  Retrieval in ~1s  Find experimentally that large vocabularies can be beneficial for recognition [Nister & Stewenius, CVPR’06]

Larger vocabularies can be advantageous… But what happens if it is too large?

Bags of features for object recognition face, flowers, building  Works pretty well for image-level classification Source: Lana Lazebnik Source credit : K. Grauman, B. Leibe

Bags of features for object recognition Caltech6 dataset bag of features Parts-and-shape model Source: Lana Lazebnik

Bags of words: pros and cons + flexible to geometry / deformations / viewpoint + compact summary of image content + provides vector representation for sets + has yielded good recognition results in practice -basic model ignores geometry – must verify afterwards, or encode via features -background and foreground mixed when bag covers whole image -interest points or sampling: no guarantee to capture object-level parts -optimal vocabulary formation remains unclear 48 Source credit : K. Grauman, B. Leibe

Summary  Local invariant features: distinctive matches possible in spite of significant view change, useful not only to provide matches for multi-view geometry, but also to find objects and scenes.  To find correspondences among detected features, measure distance between descriptors, and look for most similar patches.  Bag of words representation: quantize feature space to make discrete set of visual words  Summarize image by distribution of words  Index individual words  Inverted index: pre-compute index to enable faster search at query time Source credit : K. Grauman, B. Leibe