Local invariant features Cordelia Schmid INRIA, Grenoble.

Slides:



Advertisements
Similar presentations
Feature extraction: Corners
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP)
Interest points CSE P 576 Ali Farhadi Many slides from Steve Seitz, Larry Zitnick.
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Matching with Invariant Features
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image can be encoded.
Computational Photography
Algorithms and Applications in Computer Vision
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
1 Interest Operators Find “interesting” pieces of the image –e.g. corners, salient regions –Focus attention of algorithms –Speed up computation Many possible.
Lecture 4: Feature matching
1 Interest Operator Lectures lecture topics –Interest points 1 (Linda) interest points, descriptors, Harris corners, correlation matching –Interest points.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Feature extraction: Corners and blobs
Scale Invariant Feature Transform (SIFT)
Blob detection.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image can be encoded.
1 Interest Operators Find “interesting” pieces of the image Multiple possible uses –image matching stereo pairs tracking in videos creating panoramas –object.
Summary of Previous Lecture A homography transforms one 3d plane to another 3d plane, under perspective projections. Those planes can be camera imaging.
Machine learning & category recognition Cordelia Schmid Jakob Verbeek.
Overview Introduction to local features
Computer vision.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Lecture 06 06/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
CSE 185 Introduction to Computer Vision Local Invariant Features.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Notes on the Harris Detector
Local invariant features Cordelia Schmid INRIA, Grenoble.
Harris Corner Detector & Scale Invariant Feature Transform (SIFT)
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
Project 3 questions? Interest Points and Instance Recognition Computer Vision CS 143, Brown James Hays 10/21/11 Many slides from Kristen Grauman and.
Local features and image matching October 1 st 2015 Devi Parikh Virginia Tech Disclaimer: Many slides have been borrowed from Kristen Grauman, who may.
Local features: detection and description
Features Jan-Michael Frahm.
CS654: Digital Image Analysis
Instructor: Mircea Nicolescu Lecture 10 CS 485 / 685 Computer Vision.
CSE 185 Introduction to Computer Vision Local Invariant Features.
Keypoint extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Invariant Local Features Image content is transformed into local feature coordinates that are invariant to translation, rotation, scale, and other imaging.
776 Computer Vision Jan-Michael Frahm Spring 2012.
SIFT Scale-Invariant Feature Transform David Lowe
MAN-522: Computer Vision Edge detection Feature and blob detection
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Distinctive Image Features from Scale-Invariant Keypoints
TP12 - Local features: detection and description
Source: D. Lowe, L. Fei-Fei Recap: edge detection Source: D. Lowe, L. Fei-Fei.
Local features: detection and description May 11th, 2017
Feature description and matching
CSE 455 – Guest Lectures 3 lectures Contact Interest points 1
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Local features and image matching
SIFT keypoint detection
Lecture VI: Corner and Blob Detection
Feature descriptors and matching
Lecture 5: Feature invariance
Recognition and Matching based on local invariant features
Lecture 5: Feature invariance
Presentation transcript:

Local invariant features Cordelia Schmid INRIA, Grenoble

Local features ( ) local descriptor Several / many local descriptors per image Robust to occlusion/clutter + no object segmentation required Photometric : distinctive Invariant : to image transformations + illumination changes

Local features: interest points

Local features: Contours/segments

Local features: segmentation

Application: Matching Find corresponding locations in the image

Matching algorithm 1.Extract descriptors for each image I 1 and I 2 2.Compute similarity measure between all pairs of descriptors 3.Select couples according to different strategies 1.All matches above a threshold 2.Winner takes all 3.Cross-validation matching 4.Verify neighborhood constraints 5.Compute the global geometric relation (fundamental matrix or homography) robustly 6. Repeat matching using the global geometric relation

Selection strategies Winner takes all –The best matching pairs (with the highest score) is selected –All matches with the points and are removed Cross-validation matching –For each point in image 1 keep the best match –For each point in image 2 keep the best match –Verify the matches correspond both ways

Illustration – Matching Interest points extracted with Harris detector (~ 500 points)

Matching Interest points matched based on cross-correlation (188 pairs) Illustration – Matching

Global constraints Global constraint - Robust estimation of the fundamental matrix 99 inliers 89 outliers Illustration – Matching

Application: Panorama stitching Images courtesy of A. Zisserman.

Application: Instance-level recognition Search for particular objects and scenes in large databases …

Finding the object despite possibly large changes in scale, viewpoint, lighting and partial occlusion  requires invariant description Viewpoint Scale Lighting Occlusion Difficulties

Very large images collection  need for efficient indexing –Flickr has 2 billion photographs, more than 1 million added daily –Facebook has 15 billion images (~27 million added daily) –Large personal collections –Video collections, i.e., YouTube

Search photos on the web for particular places Find these landmarks...in these images and 1M more Applications

Take a picture of a product or advertisement  find relevant information on the web [Pixee – Milpix]

Applications Finding stolen/missing objects in a large collection …

Applications Copy detection for images and videos Search in 200h of video Query video

20 K. Grauman, B. Leibe Sony Aibo – Robotics –Recognize docking station –Communicate with visual cards –Place recognition –Loop closure in SLAM Slide credit: David Lowe Applications

Instance-level recognition: Approach Extraction of invariant image descriptors Matching descriptors between images -Matching of the query images to all images of a database -Speed-up by efficient indexing structures Geometric verification –Verification of spatial consistency for a short list

Local features - history Line segments [Lowe’87, Ayache’90] Interest points & cross correlation [Z. Zhang et al. 95] Rotation invariance with differential invariants [Schmid&Mohr’96] Scale & affine invariant detectors [Lindeberg’98, Lowe’99, Tuytelaars&VanGool’00, Mikolajczyk&Schmid’02, Matas et al.’02] Dense detectors and descriptors [Leung&Malik’99, Fei-Fei& Perona’05, Lazebnik et al.’06] Contour and region (segmentation) descriptors [Shotton et al.’05, Opelt et al.’06, Ferrari et al.’06, Leordeanu et al.’07]

Local features 1) Extraction of local features –Contours/segments –Interest points & regions –Regions by segmentation –Dense features, points on a regular grid 2) Description of local features –Dependant on the feature type –Segments  angles, length ratios –Interest points  greylevels, gradient histograms –Regions (segmentation)  texture + color distributions

Line matching Extraction de contours –Zero crossing of Laplacian –Local maxima of gradients Chain contour points (hysteresis) Extraction of line segments Description of segments –Mi-point, length, orientation, angle between pairs etc.

Experimental results – line segments images 600 x 600

Experimental results – line segments 248 / 212 line segments extracted

Experimental results – line segments 89 matched line segments - 100% correct

Experimental results – line segments 3D reconstruction

Problems of line segments Often only partial extraction –Line segments broken into parts –Missing parts Information not very discriminative –1D information –Similar for many segments Potential solutions –Pairs and triplets of segments –Interest points

Overview Harris interest points Comparison of IP: SSD, ZNCC, SIFT Scale & affine invariant interest points (student presentation) Evaluation and comparison of different detectors Region descriptors and their performance

Harris detector [Harris & Stephens’88] Based on the idea of auto-correlation Important difference in all directions => interest point

Harris detector Auto-correlation function for a point and a shift

Harris detector Auto-correlation function for a point and a shift small in all directions large in one directions large in all directions { → uniform region → contour → interest point

Harris detector

Discret shifts are avoided based on the auto-correlation matrix with first order approximation

Harris detector Auto-correlation matrix the sum can be smoothed with a Gaussian

Harris detector Auto-correlation matrix –captures the structure of the local neighborhood –measure based on eigenvalues of this matrix 2 strong eigenvalues 1 strong eigenvalue 0 eigenvalue => interest point => contour => uniform region

Interpreting the eigenvalues 1 2 “Corner” 1 and 2 are large, 1 ~ 2 ; E increases in all directions 1 and 2 are small; E is almost constant in all directions “Edge” 1 >> 2 “Edge” 2 >> 1 “Flat” region Classification of image points using eigenvalues of M:

Corner response function “Corner” R > 0 “Edge” R < 0 “Flat” region |R| small α: constant (0.04 to 0.06)

Harris detector Reduces the effect of a strong contour Cornerness function Interest point detection –Treshold (absolut, relatif, number of corners) –Local maxima

Harris Detector: Steps

Compute corner response R

Harris Detector: Steps Find points with large corner response: R>threshold

Harris Detector: Steps Take only the points of local maxima of R

Harris Detector: Steps

Harris detector: Summary of steps 1.Compute Gaussian derivatives at each pixel 2.Compute second moment matrix M in a Gaussian window around each pixel 3.Compute corner response function R 4.Threshold R 5.Find local maxima of response function (non-maximum suppression)

Harris - invariance to transformations Geometric transformations –translation –rotation –similitude (rotation + scale change) –affine (valide for local planar objects) Photometric transformations –Affine intensity changes (I  a I + b)

Harris Detector: Invariance Properties Rotation Ellipse rotates but its shape (i.e. eigenvalues) remains the same Corner response R is invariant to image rotation

Harris Detector: Invariance Properties Affine intensity change Only derivatives are used => invariance to intensity shift I  I + b Intensity scale: I  a I R x (image coordinate) threshold R x (image coordinate) Partially invariant to affine intensity change, dependent on type of threshold

Harris Detector: Invariance Properties Scaling All points will be classified as edges Corner Not invariant to scaling

Overview Harris interest points Comparison of IP: SSD, ZNCC, SIFT Scale & affine invariant interest points (student presentation) Evaluation and comparison of different detectors Region descriptors and their performance

Comparison of patches - SSD image 1 image 2 SSD : sum of square difference Comparison of the intensities in the neighborhood of two interest points Small difference values  similar patches

Comparison of patches SSD : Invariance to photometric transformations? Intensity changes (I  I + b) Intensity changes (I  aI + b) => Normalizing with the mean of each patch => Normalizing with the mean and standard deviation of each patch

Cross-correlation ZNCC ZNCC: zero normalized cross correlation zero normalized SSD ZNCC values between -1 and 1, 1 when identical patches in practice threshold around 0.5

Local descriptors Greyvalue derivatives Differential invariants [Koenderink’87] SIFT descriptor [Lowe’99]

The gradient points in the direction of most rapid increase in intensity Greyvalue derivatives: Image gradient The gradient of an image: The gradient direction is given by –how does this relate to the direction of the edge? The edge strength is given by the gradient magnitude Source: Steve Seitz

Differentiation and convolution Recall, for 2D function, f(x,y): We could approximate this as 1 Source: D. Forsyth, D. Lowe Convolution with the filter

Finite difference filters Other approximations of derivative filters exist: Source: K. Grauman

Effects of noise Consider a single row or column of the image –Plotting intensity as a function of position gives a signal Where is the edge? Source: S. Seitz

Solution: smooth first To find edges, look for peaks in f g f * g Source: S. Seitz

Differentiation is convolution, and convolution is associative: This saves us one operation: Derivative theorem of convolution f Source: S. Seitz

Local descriptors Greyvalue derivatives –Convolution with Gaussian derivatives

Local descriptors Notation for greyvalue derivatives [Koenderink’87] Invariance?

Local descriptors – rotation invariance Invariance to image rotation : differential invariants [Koen87] gradient magnitude Laplacian

Laplacian of Gaussian (LOG)

Local descriptors – rotation invariance Gaussian derivative-based descriptors –Steerable filters (Freeman and Adelson’91) –“Steering the derivatives in the direction of an angle “

SIFT descriptor [Lowe’99] Approach –8 orientations of the gradient –4x4 spatial grid –soft-assignment to spatial bins, dimension 128 –normalization of the descriptor to norm one –comparison with Euclidean distance gradient 3D histogram   image patch y x

Local descriptors - rotation invariance Estimation of the dominant orientation –extract gradient orientation –histogram over gradient orientation –peak in this histogram Rotate patch in dominant direction 0 2 

Local descriptors – illumination change in case of an affine transformation Robustness to illumination changes

Local descriptors – illumination change in case of an affine transformation Normalization of derivatives with gradient magnitude Robustness to illumination changes

Local descriptors – illumination change in case of an affine transformation Normalization of the image patch with mean and variance Normalization of derivatives with gradient magnitude Robustness to illumination changes

Invariance to scale changes Scale change between two images Scale factor s can be eliminated Support region for calculation!! –In case of a convolution with Gaussian derivatives defined by

Scale invariance - motivation Description regions have to be adapted to scale changes Interest points have to be repeatable for scale changes

Harris detector + scale changes Repeatability rate