SIFT keypoint detection

Slides:



Advertisements
Similar presentations
Feature Detection. Description Localization More Points Robust to occlusion Works with less texture More Repeatable Robust detection Precise localization.
Advertisements

Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
CSE 473/573 Computer Vision and Image Processing (CVIP)
Interest points CSE P 576 Ali Farhadi Many slides from Steve Seitz, Larry Zitnick.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
CSE 473/573 Computer Vision and Image Processing (CVIP)
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Matching with Invariant Features
Algorithms and Applications in Computer Vision
Interest points CSE P 576 Larry Zitnick Many slides courtesy of Steve Seitz.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Lecture 6: Feature matching CS4670: Computer Vision Noah Snavely.
Lecture 4: Feature matching
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Distinctive Image Feature from Scale-Invariant KeyPoints
Feature extraction: Corners and blobs
Scale Invariant Feature Transform (SIFT)
Blob detection.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
CS4670: Computer Vision Kavita Bala Lecture 8: Scale invariance.
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe – IJCV 2004 Brien Flewelling CPSC 643 Presentation 1.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Overview Introduction to local features
Computer vision.
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
CS 558 C OMPUTER V ISION Lecture VII: Corner and Blob Detection Slides adapted from S. Lazebnik.
Lecture 06 06/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Lecture 4: Feature matching CS4670 / 5670: Computer Vision Noah Snavely.
CS 1699: Intro to Computer Vision Local Image Features: Extraction and Description Prof. Adriana Kovashka University of Pittsburgh September 17, 2015.
CVPR 2003 Tutorial Recognition and Matching Based on Local Invariant Features David Lowe Computer Science Department University of British Columbia.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Harris Corner Detector & Scale Invariant Feature Transform (SIFT)
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
Project 3 questions? Interest Points and Instance Recognition Computer Vision CS 143, Brown James Hays 10/21/11 Many slides from Kristen Grauman and.
Local features: detection and description
CS654: Digital Image Analysis
Presented by David Lee 3/20/2006
CSE 185 Introduction to Computer Vision Local Invariant Features.
Keypoint extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Blob detection.
776 Computer Vision Jan-Michael Frahm Spring 2012.
MAN-522: Computer Vision Edge detection Feature and blob detection
Interest Points EE/CSE 576 Linda Shapiro.
Lecture 13: Feature Descriptors and Matching
Presented by David Lee 3/20/2006
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Distinctive Image Features from Scale-Invariant Keypoints
3D Vision Interest Points.
TP12 - Local features: detection and description
Scale and interest point descriptors
Local features: detection and description May 11th, 2017
Feature description and matching
Corners and Interest Points
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Interest Points & Descriptors 3 - SIFT
CSE 185 Introduction to Computer Vision
Lecture VI: Corner and Blob Detection
Feature descriptors and matching
Lecture 5: Feature invariance
Presented by Xu Miao April 20, 2005
Lecture 5: Feature invariance
Presentation transcript:

SIFT keypoint detection D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp. 91-110, 2004

Keypoint detection with scale selection We want to extract keypoints with characteristic scales that are covariant w.r.t. the image transformation

Basic idea Convolve the image with a “blob filter” at multiple scales and look for extrema of filter response in the resulting scale space T. Lindeberg, Feature detection with automatic scale selection, IJCV 30(2), pp 77-116, 1998

* = Blob detection minima maxima Find maxima and minima of blob filter response in space and scale Source: N. Snavely

Blob filter Laplacian of Gaussian: Circularly symmetric operator for blob detection in 2D

Recall: Edge detection f Derivative of Gaussian Edge = maximum of derivative Source: S. Seitz

Edge detection, Take 2 f Edge Second derivative of Gaussian (Laplacian) An edge filtered with the Laplacian of Gaussian is the difference between the blurry edge and the original edge. Edge = zero crossing of second derivative Source: S. Seitz

From edges to blobs Edge = ripple Blob = superposition of two ripples maximum Spatial selection: the magnitude of the Laplacian response will achieve a maximum at the center of the blob, provided the scale of the Laplacian is “matched” to the scale of the blob

original signal (radius=8) Scale selection We want to find the characteristic scale of the blob by convolving it with Laplacians at several scales and looking for the maximum response However, Laplacian response decays as scale increases: increasing σ original signal (radius=8)

Scale normalization The response of a derivative of Gaussian filter to a perfect step edge decreases as σ increases: To keep response the same (scale-invariant), must multiply Gaussian derivative by σ Laplacian is the second Gaussian derivative, so it must be multiplied by σ2 The area under the first derivative of Gaussian from –infinity to 0 is equal to the value of the Gaussian at zero

Effect of scale normalization Original signal Unnormalized Laplacian response Scale-normalized Laplacian response maximum

Blob detection in 2D Scale-normalized Laplacian of Gaussian:

Blob detection in 2D At what scale does the Laplacian achieve a maximum response to a binary circle of radius r? r image Laplacian

Blob detection in 2D At what scale does the Laplacian achieve a maximum response to a binary circle of radius r? To get maximum response, the zeros of the Laplacian have to be aligned with the circle The Laplacian is given by (up to scale): Therefore, the maximum response occurs at circle Laplacian r image

Scale-space blob detector Convolve image with scale-normalized Laplacian at several scales

Scale-space blob detector: Example

Scale-space blob detector: Example

Scale-space blob detector Convolve image with scale-normalized Laplacian at several scales Find maxima of squared Laplacian response in scale-space

Scale-space blob detector: Example

Efficient implementation Approximating the Laplacian with a difference of Gaussians: (Laplacian) (Difference of Gaussians)

Efficient implementation David G. Lowe. "Distinctive image features from scale-invariant keypoints.” IJCV 60 (2), pp. 91-110, 2004.

Eliminating edge responses Laplacian has strong response along edges

Eliminating edge responses Laplacian has strong response along edges Solution: filter based on Harris response function over neighboroods containing the “blobs”

From feature detection to feature description To recognize the same pattern in multiple images, we need to match appearance “signatures” in the neighborhoods of extracted keypoints But corresponding neighborhoods can be related by a scale change or rotation We want to normalize neighborhoods to make signatures invariant to these transformations

Finding a reference orientation Create histogram of local gradient directions in the patch Assign reference orientation at peak of smoothed histogram 2 p

SIFT features Detected features with characteristic scales and orientations: David G. Lowe. "Distinctive image features from scale-invariant keypoints.” IJCV 60 (2), pp. 91-110, 2004.

From keypoint detection to feature description Detection is covariant: features(transform(image)) = transform(features(image)) Description is invariant: features(transform(image)) = features(image)

SIFT descriptors Inspiration: complex neurons in the primary visual cortex From the paper: The previous operations have assigned an image location, scale, and orientation to each keypoint. These parameters impose a repeatable local 2D coordinate system in which to describe the local image region, and therefore provide invariance to these parameters. The next step is to compute a descriptor for the local image region that is highly distinctive yet is as invariant as possible to remaining variations, such as change in illumination or 3D viewpoint. One obvious approach would be to sample the local image intensities around the keypoint at the appropriate scale, and to match these using a normalized correlation measure. However, simple correlation of image patches is highly sensitive to changes that cause misregistration of samples, such as affine or 3D viewpoint change or non-rigid deformations. A better approach has been demonstrated by Edelman, Intrator, and Poggio (1997). Their proposed representation was based upon a model of biological vision, in particular of complex neurons in primary visual cortex. These complex neurons respond to a gradient at a particular orientation and spatial frequency, but the location of the gradient on the retina is allowed to shift over a small receptive field rather than being precisely localized. Edelman et al. hypothesized that the function of these complex neurons was to allow for matching and recognition of 3D objects from a range of viewpoints. They have performed detailed experiments using 3D computer models of object and animal shapes which show that matching gradients while allowing for shifts in their position results in much better classification under 3D rotation. For example, recognition accuracy for 3D objects rotated in depth by 20 degrees increased from 35% for correlation of gradients to 94% using the complex cell model. Our implementation described below was inspired by this idea, but allows for positional shift using a different computational mechanism. D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), pp. 91-110, 2004

Properties of SIFT Extraordinarily robust detection and description technique Can handle changes in viewpoint Up to about 60 degree out-of-plane rotation Can handle significant changes in illumination Sometimes even day vs. night Fast and efficient—can run in real time Lots of code available Source: N. Snavely

A hard keypoint matching problem NASA Mars Rover images

Answer below (look for tiny colored squares…) NASA Mars Rover images with SIFT feature matches Figure by Noah Snavely

What about 3D rotations?

What about 3D rotations? Affine transformation approximates viewpoint changes for roughly planar objects and roughly orthographic cameras

Affine adaptation Consider the second moment matrix of the window containing the blob: direction of the slowest change direction of the fastest change (max)-1/2 (min)-1/2 Recall: This ellipse visualizes the “characteristic shape” of the window

Affine adaptation K. Mikolajczyk and C. Schmid, Scale and affine invariant interest point detectors, IJCV 60(1):63-86, 2004

Keypoint detectors/descriptors for recognition: A retrospective Detected features S. Lazebnik, C. Schmid, and J. Ponce, A Sparse Texture Representation Using Affine-Invariant Regions, CVPR 2003

Keypoint detectors/descriptors for recognition: A retrospective Detected features R. Fergus, P. Perona, and A. Zisserman, Object Class Recognition by Unsupervised Scale-Invariant Learning, CVPR 2003 – winner of 2013 Longuet-Higgins Prize

Keypoint detectors/descriptors for recognition: A retrospective S. Lazebnik, C. Schmid, and J. Ponce, Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories, CVPR 2006 – winner of 2016 Longuet-Higgins Prize

Keypoint detectors/descriptors for recognition: A retrospective level 2 level 0 level 1 S. Lazebnik, C. Schmid, and J. Ponce, Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories, CVPR 2006 – winner of 2016 Longuet-Higgins Prize

Keypoint detectors/descriptors for recognition: A retrospective