Top-points in Image Matching Bram Platel Evgenya Balmashnova Luc Florack Bart ter Haar Romeny.

Slides:



Advertisements
Similar presentations
Ter Haar Romeny, FEV Geometry-driven diffusion: nonlinear scale-space – adaptive scale-space.
Advertisements

Ter Haar Romeny, ICPR 2010 Introduction to Scale-Space and Deep Structure.
Active Appearance Models
Scale & Affine Invariant Interest Point Detectors Mikolajczyk & Schmid presented by Dustin Lennon.
Presented by Xinyu Chang
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Investigation Into Optical Flow Problem in the Presence of Spatially-varying Motion Blur Mohammad Hossein Daraei June 2014 University.
Ter Haar Romeny, Computer Vision 2014 Geometry-driven diffusion: nonlinear scale-space – adaptive scale-space.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
The Image Histogram.
Distinctive Image Features from Scale- Invariant Keypoints Mohammad-Amin Ahantab Technische Universität München, Germany.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
La Parguera Hyperspectral Image size (250x239x118) using Hyperion sensor. INTEREST POINTS FOR HYPERSPECTRAL IMAGES Amit Mukherjee 1, Badrinath Roysam 1,
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Lecture 4: Feature matching
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
Feature extraction: Corners and blobs
Scale Invariant Feature Transform (SIFT)
Blob detection.
Edge Detection Today’s readings Cipolla and Gee –supplemental: Forsyth, chapter 9Forsyth Watt, From Sandlot ScienceSandlot Science.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
CS4670: Computer Vision Kavita Bala Lecture 7: Harris Corner Detection.
CS4670: Computer Vision Kavita Bala Lecture 8: Scale invariance.
Lecture 6: Feature matching and alignment CS4670: Computer Vision Noah Snavely.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Overview Introduction to local features
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Computer vision.
05 - Feature Detection Overview Feature Detection –Intensity Extrema –Blob Detection –Corner Detection Feature Descriptors Feature Matching Conclusion.
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
Edge Linking & Boundary Detection
CVPR 2003 Tutorial Recognition and Matching Based on Local Invariant Features David Lowe Computer Science Department University of British Columbia.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
CSCE 643 Computer Vision: Extractions of Image Features Jinxiang Chai.
CS654: Digital Image Analysis Lecture 25: Hough Transform Slide credits: Guillermo Sapiro, Mubarak Shah, Derek Hoiem.
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Local invariant features Cordelia Schmid INRIA, Grenoble.
Bart M. ter Haar Romeny.  Question: can top-points be used for object- retrieval tasks?
Harris Corner Detector & Scale Invariant Feature Transform (SIFT)
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
CSE 185 Introduction to Computer Vision Feature Matching.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
3D Reconstruction Using Image Sequence
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Introduction to Scale Space and Deep Structure. Importance of Scale Painting by Dali Objects exist at certain ranges of scale. It is not known a priory.
Instructor: Mircea Nicolescu Lecture 7
Instructor: Mircea Nicolescu Lecture 5 CS 485 / 685 Computer Vision.
Digital Image Processing CSC331
Instructor: Mircea Nicolescu Lecture 10 CS 485 / 685 Computer Vision.
Distinctive Image Features from Scale-Invariant Keypoints Presenter :JIA-HONG,DONG Advisor : Yen- Ting, Chen 1 David G. Lowe International Journal of Computer.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Image Representation and Description – Representation Schemes
SIFT Scale-Invariant Feature Transform David Lowe
CS262: Computer Vision Lect 09: SIFT Descriptors
Miguel Tavares Coimbra
Feature description and matching
CAP 5415 Computer Vision Fall 2012 Dr. Mubarak Shah Lecture-5
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
ECE734 Project-Scale Invariant Feature Transform Algorithm
Presentation transcript:

Top-points in Image Matching Bram Platel Evgenya Balmashnova Luc Florack Bart ter Haar Romeny

Introduction Question: can top-points be used for object-retrieval tasks?

Introduction to Scale Space and Deep Structure

Importance of Scale in Image Analysis Painting by Dali Objects in images exist at different ranges of scale. Usually it is not known a priory at what scale to look.

At the original scale of a dithered image we cannot calculate a derivative. We need to observe the image at a certain scale. BLUR

Solution? Look at all scales simultaneously Scale x y Scale Space

Scale Space in Human Vision The human visual system is a multi-scale sampling device The retina contains receptive fields; groups of receptors assembled in such a way that they form a set of apertures of widely varying size.

Practical Implementation Convolve the image with a Gaussian Kernel

We can Calculate Derivatives and Combinations of them at all Scales Gradient Magnitude Laplacian Original Image

Main Topic In this presentation we will show how we can exploit the deep structure of images to define invariant interest points and features which can be used for matching problems in computer vision. We consider only grey-value images.

Interest Points The locations of particularly characteristic points are called the interest points or key points. These interest points have to be as invariant as possible, but at the same time they have to carry a lot of distinctive information.

Why Interest Points in Scale Space? Information in interest points is defined by their neighborhood. But how big should we choose this neighborhood? Let’s take the corners of the mouth as interest points. The red circles are the areas in which the information is gathered. If we make the picture bigger, the size of the neighborhood is too small. The neighborhood should scale with the image

Why Interest Points in Scale Space? When the interest points are detected in scale space they do not only have spatial coordinates x and y, but also a scale . This scale tells us how big the neighborhood should be.  

Which Interest Points to Use? Our interest points have to be detected in scale space. They also have to… –…contain a lot of information –…be reproducible –…be stable –…be well understood

We Suggest Top-Points The points we introduce have these desired properties.

Critical Points, Paths and Top-Points Maxima Minimum Saddles  L=0 Critical Points

Critical Points, Paths and Top-Points Maxima Minimum Saddles  L=0 Critical Points Det(H)=0 Top-Points

Possible to calculate them for every Function of the Image L(x,y,  ) OriginalGradient Magnitude LaplacianDet(H)

Detecting Critical Paths Since for a critical path  L=0 Intersection of Level Surfaces L x =0 with L y =0 Will give the critical paths.

Detecting Top-Points Since for a top-point both  L=0 and Det[H]=Lxx Lyy- Lxy 2 =0 We can find them by intersecting the paths with the level surface Det[H]=0

Invariance of Top Points Top-points are invariant to certain transformations. By invariant we mean that they move according to the transformation. Allowed Trans.

Reconstruction It is possible to make a reconstruction of the original image from its top-points. We can generate reconstructed images which give the same (plus more) top- points as the original image. This reconstruction resembles the original image.

Original Image Top-Points and Features Reconstruction

Metameric Class Original By adjusting boundary and smoothness constraints we can improve the visual performance. For this 300x300 picture 1000 top-points with 6 features were used.

This tells us that the top-points indeed contain a lot of information about the image.

Localization of Top-Points For points close to top-points it is possible to calculate a vector pointing towards the position of the top-point. x y  Approximated Top- Points Displacement Vectors Real Locations

Localization of Top-Points For points close to top-points it is possible to calculate a vector pointing towards the position of the top-point. This enables us to use fast top-point detection algorithms which do not have to be very accurate.

Stability of Top-Points The locations of top-points change when noise is added to the image.

Stability of Top-Points We can calculate the variance of the displacement of top- points under noise. We need 4 th order derivatives in the top-points for that.

Selecting Stable Paths Large instabilities are found in flat areas in the image. Areas with a lot of structure result in more stable top points. In the article we used an adapted version of the TV-norm (Total Variation) to measure the flatness around a top point. We can give top points a weight in the EMD algorithm based on the stability of the point.

Calculating the Stability Norm Taylor Expansion (2 nd order) Limiting Procedure Koenderink’s deviation from flatness xcxc x  Integration Area

Thresholding on Stability Stable Paths Unstable Paths

Database Image Retrieval Databas e Query Image

Experiments A simple image retrieval task. Using a small version of the Olivetti Faces Database. Consisting of 200 images of 20 different people (10 p.p.)

The Deep Structure of an Image Look at all scales simultaneously Scale x y

Critical Points, Paths and Top Points Maxima Minimum Saddles  u=0 Critical Points Det(H)=0 Top Points

Comparing Top Points of Images CompareEMD

Earth Movers Distance (EMD) wiwi f ij c ij AB ujuj [*]Rubner, Tomasi, Guibas, 1998, IEEE Conf. on Computer Vision PilesHoles

Calculating the EMD To calculate the earth movers distance we need: –Weights for our top points (e.g. more important points could contain more weight). –A cost function for transporting the weights between point sets (we incorporated a distance function between top points). MMA

Measuring Distance in Scale Space Scale Space is not a Euclidean Space. The simple Euclidean Metric no longer holds. Scale x y

Eberly’s Distance Measure in Scale Space For an infinitely small step in scale space the following distance function holds: Given this distance function we can find a unique geodesic path connecting two points in scale space. The distance between two points is measured along the this geodesic path connecting these points.  0 relates the importance of scale- to spatial measurements

Distance Between Points in Scale Space By parameterizing along the geodesic path the distance function can be derived by solving an ordinary differential equation. This yields the following equation describing the distance between two points in scale space:

We now have a distance measure which we can use in our EMD algorithm. We have introduced a tunable parameter  (>0). We now need to give a weight to our points.

Results a93%78%68%62%58%54%49%45%43% b93%82%90%73%70%68%63%59%56% c95%88%83%76%73%66%62%59%58% d100 % 97%96%95%92%88%86%84%81% a.Using Euclidean Distance b.Using Eberly Distance c.As b. including stability norm d.As c. including 2 nd order derivatives.

Distinctive Features To distinguish top-points from each other a set of distinctive features are needed in every top-point. These local features describe the neighborhood of the top-point.

Differential Invariants We use the complete set of irreducible 3 rd order differential invariants. These features are rotation and scaling invariant.

Matching With the top-points as interest points and the differential invariants as the descriptors we can now start matching them across images.

The Task We have a scene and from that scene we want to retrieve the location of the query object.

The first step of the matching algorithm is to find the location of our highly invariant interest points, the top-points.

The top-points and differential invariants are calculated for the query object and the scene.

The location in scale-space (x,y,  ), the gradient angle  and the differential invariant features (D1,D2,D3,D4,D5,D6) are stored for every top-point.

Distance between Feature Vectors A sensible distance between feature vectors is essential. We have used Euclidean distance on ‘normalized’ differential invariants. We tried Mahalanobis distance obtained from a training set.

Drawbacks of these Methods Euclidean distance is not optimal since the normalization of the features still does not make them Euclidean. The covariance matrix C of the Mahalanobis distance is global and not defined for each feature vector and training is required.

Similarity Measure We can calculate the propagation of noise in scale space* This enables us to calculate a covariance matrix  for each feature vector. The dissimilarity (“distance”) measure is expressed as: *Topological and Geometrical Aspects of Image Structure, Johan Blom

Improvement Improvement e.g. for 5% noise addition. True Positive Rate False Positive Rate

ROC curves

We now compare the differential invariant features. compare distance = 0.5distance = 0.2distance = 0.3

The vectors with the smallest distance are paired. smallest distance distance = 0.2

A set of coordinates is formed from the differences in scale (Log(  o1 )- Log(  s2 )) and in angles (  o1 -  s2 ). (  ,   )

  Important Clusters For these clusters we calculate the mean  and  Clustering ( ,  ) If these coordinates are plotted in a scatter plot clusters can be identified. In this scatter plot we find two dense clusters

The stability criterion removes much of the scatter

After finding  and  for each cluster we transform the (x,y,  ) parameters of the object top-points to match the scene.

Rotate and scale according to the cluster means.

After this step we can find the translation of the points in scene. By again doing the clustering but now for coordinates (  x,  y).

The translations we find correspond to the location of the objects in the scene.

In this example we have two clusters of correctly matched points. C1 C2

The transformation of each object in the scene matching to the query object is known from the clustering.

We can transform the outline of the query object and project it on the scene image.

Object Retrieval Results

Invariance to Sampling

Conclusion Top-points have proved to be invariant interest points which are useful for matching. The differential invariants have shown to be very distinctive. Experiments show good results.