CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

Feature Detection. Description Localization More Points Robust to occlusion Works with less texture More Repeatable Robust detection Precise localization.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
Presented by Xinyu Chang
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
CSE 473/573 Computer Vision and Image Processing (CVIP)
Distinctive Image Features from Scale- Invariant Keypoints Mohammad-Amin Ahantab Technische Universität München, Germany.
IBBT – Ugent – Telin – IPI Dimitri Van Cauwelaert A study of the 2D - SIFT algorithm Dimitri Van Cauwelaert.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Fast High-Dimensional Feature Matching for Object Recognition David Lowe Computer Science Department University of British Columbia.
Lecture 6: Feature matching CS4670: Computer Vision Noah Snavely.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Distinctive Image Feature from Scale-Invariant KeyPoints
Feature extraction: Corners and blobs
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
Scale Invariant Feature Transform (SIFT)
Automatic Matching of Multi-View Images
SIFT - The Scale Invariant Feature Transform Distinctive image features from scale-invariant keypoints. David G. Lowe, International Journal of Computer.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe – IJCV 2004 Brien Flewelling CPSC 643 Presentation 1.
Sebastian Thrun CS223B Computer Vision, Winter Stanford CS223B Computer Vision, Winter 2005 Lecture 3 Advanced Features Sebastian Thrun, Stanford.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Interest Point Descriptors
Distinctive Image Features from Scale-Invariant Keypoints By David G. Lowe, University of British Columbia Presented by: Tim Havinga, Joël van Neerbos.
Computer vision.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
Bag of Visual Words for Image Representation & Visual Search Jianping Fan Dept of Computer Science UNC-Charlotte.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
CVPR 2003 Tutorial Recognition and Matching Based on Local Invariant Features David Lowe Computer Science Department University of British Columbia.
CSCE 643 Computer Vision: Extractions of Image Features Jinxiang Chai.
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Distinctive Image Features from Scale-Invariant Keypoints Ronnie Bajwa Sameer Pawar * * Adapted from slides found online by Michael Kowalski, Lehigh University.
Harris Corner Detector & Scale Invariant Feature Transform (SIFT)
Distinctive Image Features from Scale-Invariant Keypoints David Lowe Presented by Tony X. Han March 11, 2008.
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Local features: detection and description
Scale Invariant Feature Transform (SIFT)
Distinctive Image Features from Scale-Invariant Keypoints Presenter :JIA-HONG,DONG Advisor : Yen- Ting, Chen 1 David G. Lowe International Journal of Computer.
Blob detection.
SIFT.
SIFT Scale-Invariant Feature Transform David Lowe
CS262: Computer Vision Lect 09: SIFT Descriptors
Interest Points EE/CSE 576 Linda Shapiro.
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Distinctive Image Features from Scale-Invariant Keypoints
Scale Invariant Feature Transform (SIFT)
SIFT paper.
TP12 - Local features: detection and description
Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2006/3/22
Scale and interest point descriptors
From a presentation by Jimmy Huff Modified by Josiah Yoder
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Interest Points & Descriptors 3 - SIFT
SIFT.
Edge detection f(x,y) viewed as a smooth function
ECE734 Project-Scale Invariant Feature Transform Algorithm
Lecture VI: Corner and Blob Detection
Lecture 5: Feature invariance
SIFT SIFT is an carefully designed procedure with empirically determined parameters for the invariant and distinctive features.
Presented by Xu Miao April 20, 2005
Presentation transcript:

CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5 Univ. of Central Florida Office 247-F HEC Lecture-5

SIFT: David Lowe, UBC

SIFT - Key Point Extraction Stands for scale invariant feature transform Patented by university of British Columbia Similar to the one used in primate visual system (human, ape, monkey, etc.) Transforms image data into scale- invariant coordinates D. Lowe. Distinctive image features from scale-invariant key points., International Journal of Computer Vision 2004.

Goal Extracting distinctive invariant features Correctly matched against a large database of features from many images Invariance to image scale and rotation Robustness to Affine distortion, Change in 3D viewpoint, Addition of noise, Change in illumination.

Advantages Locality: features are local, so robust to occlusion and clutter Distinctiveness: individual features can be matched to a large database of objects Quantity: many features can be generated for even small objects Efficiency: close to real-time performance

Invariant Local Features

Steps for Extracting Key Points Scale space peak selection Potential locations for finding features Key point localization Accurately locating the feature key points Orientation Assignment Assigning orientation to the key points Key point descriptor Describing the key point as a high dimensional vector

Building a Scale Space All scales must be examined to identify scale- invariant features An efficient function is to compute the Laplacian Pyramid (Difference of Gaussian) (Burt & Adelson, 1983)

Approximation of LoG by Difference of Gaussians G  2G Heat Equation  2G  G  G(x, y, k )  G(x, y, )  k   G(x, y, k )  G(x, y, )  (k 1) 22G Typical values:  1.6; k  2

Building a Scale Space e( x2  y 2 ) / 2 k 2 2 k  2 1 G( x, y,k )  k  2

Building a Scale Space e( x2  y 2 ) / 2 k 2 2 k  2 1 G( x, y,k )  k 4 k 3 k 2 k 4 k 3 k 2 k  k  2

  .707187.6; k  2

How many scales per octave?

Initial value of sigma

Scale Space Peak Detection Compare a pixel (X) with 26 pixels in current and adjacent scales (Green Circles) Select a pixel (X) if larger/smaller than all 26 pixels Large number of extrema, computationally expensive Detect the most stable subset with a coarse sampling of scales

Key Point Localization Candidates are chosen from extrema detection original image extrema locations

Initial Outlier Rejection Low contrast candidates Poorly localized candidates along an edge Taylor series expansion of DOG, D. Minima or maxima is located at Value of D(x) at minima/maxima must be large, |D(x)|>th.

Initial Outlier Rejection from 832 key points to 729 key points, th=0.03.

Further Outlier Rejection DOG has strong response along edge Assume DOG as a surface Compute principal curvatures (PC) Along the edge one of the PC is very low, across the edge is high PC gradients along x and y directions

Further Outlier Rejection Analogous to Harris corner detector Compute Hessian of D Tr(H )  Dxx  Dyy 12 Det(H )  D D  (D )  2   xx yy xy 1 2 Remove outliers by evaluating Tr(H )2  (r 1)2 Det(H ) r r  1 2

Further Outlier Rejection Following quantity is minimum when r=1 It increases with r  r  1 Tr(H )2  (r 1)2 Det(H ) r Eliminate key points if 2 r  10

Further Outlier Rejection from 729 key points to 536 key points.

Orientation Assignment To achieve rotation invariance Compute central derivatives, gradient magnitude and direction of L (smooth image) at the scale of key point (x,y)

Orientation Assignment Create a weighted direction histogram in a neighborhood of a key point (36 bins) Weights are Gradient magnitudes Spatial gaussian filter with =1.5 x <scale of key point>

Orientation Assignment Select the peak as direction of the key point Introduce additional key points (same location) at local peaks (within 80% of max peak) of the histogram with different directions

Extraction of Local Image Descriptors at Key Points Compute relative orientation and magnitude in a 16x16 neighborhood at key point Form weighted histogram (8 bin) for 4x4 regions Weight by magnitude and spatial Gaussian Concatenate 16 histograms in one long vector of 128 dimensions 16 histograms x 8 orientations = 128 features

Descriptor Regions (n by n)

Key point matching Match the key points against a database of that obtained from training images. Find the nearest neighbor i.e. a key point with minimum Euclidean distance. Efficient Nearest Neighbor matching Looks at ratio of distance between best and 2nd best match (.8) 43

Matching local features Kristen Grauman

Matching local features ? Image 1 Image 2 To generate candidate matches, find patches that have the most similar appearance or SIFT descriptor Simplest approach: compare them all, take the closest (or closest k, or within a thresholded distance) Kristen Grauman

? ? ? ? Ambiguous matches At what distance do we have a good match? Image 1 At what distance do we have a good match? To add robustness to matching, can consider ratio : distance to best match / distance to second best match If low, first match looks good. If high, could be ambiguous match. Image 2 Kristen Grauman

The ratio of distance from the closest to the distance of the second closest

Reference D. Lowe. Distinctive image features from scale- invariant key points., International Journal of Computer Vision 2004.