Feature Reconstruction Using Lucas-Kanade Feature Tracking and Tomasi-Kanade Factorization EE7740 Project I Dr. Gunturk.

Slides:



Advertisements
Similar presentations
Feature Detection. Description Localization More Points Robust to occlusion Works with less texture More Repeatable Robust detection Precise localization.
Advertisements

The fundamental matrix F
Investigation Into Optical Flow Problem in the Presence of Spatially-varying Motion Blur Mohammad Hossein Daraei June 2014 University.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Computer Vision Optical Flow
Structure from motion.
Motion Estimation I What affects the induced image motion? Camera motion Object motion Scene structure.
LUCAS KANADE FEATURE TRACKER a pyramidal implementation
Announcements Quiz Thursday Quiz Review Tomorrow: AV Williams 4424, 4pm. Practice Quiz handout.
Today Feature Tracking Structure from Motion on Monday (1/29)
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Announcements Project1 artifact reminder counts towards your grade Demos this Thursday, 12-2:30 sign up! Extra office hours this week David (T 12-1, W/F.
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Feature extraction: Corners and blobs
CSc83029 – 3-D Computer Vision/ Ioannis Stamos 3-D Computational Vision CSc Optical Flow & Motion The Factorization Method.
Announcements Project 1 test the turn-in procedure this week (make sure your folder’s there) grading session next Thursday 2:30-5pm –10 minute slot to.
Optical flow and Tracking CISC 649/849 Spring 2009 University of Delaware.
Motion Computing in Image Analysis
Optical Flow Estimation
Lecture 19: Optical flow CS6670: Computer Vision Noah Snavely
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Numerical Recipes (Newton-Raphson), 9.4 (first.
1 Stanford CS223B Computer Vision, Winter 2006 Lecture 7 Optical Flow Professor Sebastian Thrun CAs: Dan Maynes-Aminzade, Mitul Saha, Greg Corrado Slides.
COMP 290 Computer Vision - Spring Motion II - Estimation of Motion field / 3-D construction from motion Yongjik Kim.
Image Primitives and Correspondence
3D Rigid/Nonrigid RegistrationRegistration 1)Known features, correspondences, transformation model – feature basedfeature based 2)Specific motion type,
Matching Compare region of image to region of image. –We talked about this for stereo. –Important for motion. Epipolar constraint unknown. But motion small.
KLT tracker & triangulation Class 6 Read Shi and Tomasi’s paper on good features to track
Optical Flow Digital Photography CSE558, Spring 2003 Richard Szeliski (notes cribbed from P. Anandan)
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Announcements Project1 due Tuesday. Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Supplemental:
CSCE 641 Computer Graphics: Image Registration Jinxiang Chai.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
A plane-plus-parallax algorithm Basic Model: When FOV is not very large and the camera motion has a small rotation, the 2D displacement (u,v) of an image.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Optical Flow Donald Tanguay June 12, Outline Description of optical flow General techniques Specific methods –Horn and Schunck (regularization)
CSCE 643 Computer Vision: Structure from Motion
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
CSE 185 Introduction to Computer Vision Feature Tracking and Optical Flow.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Pyramidal Implementation of Lucas Kanade Feature Tracker Jia Huang Xiaoyan Liu Han Xin Yizhen Tan.
3D Imaging Motion.
Joint Tracking of Features and Edges STAN BIRCHFIELD AND SHRINIVAS PUNDLIK CLEMSON UNIVERSITY ABSTRACT LUCAS-KANADE AND HORN-SCHUNCK JOINT TRACKING OF.
Miguel Tavares Coimbra
Feature extraction: Corners and blobs. Why extract features? Motivation: panorama stitching We have two images – how do we combine them?
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
Motion Estimation I What affects the induced image motion?
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Structure from Motion ECE 847: Digital Image Processing
Structure from Motion Paul Heckbert, Nov , Image-Based Modeling and Rendering.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
Motion / Optical Flow II Estimation of Motion Field Avneesh Sud.
Motion estimation Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/4/12 with slides by Michael Black and P. Anandan.
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Instructor: Mircea Nicolescu Lecture 10 CS 485 / 685 Computer Vision.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Motion estimation Parametric motion (image alignment) Tracking Optical flow.
Motion estimation Digital Visual Effects, Spring 2005 Yung-Yu Chuang 2005/3/23 with slides by Michael Black and P. Anandan.
CPSC 641: Image Registration
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
3D Vision Interest Points.
Motion and Optical Flow
Motion Estimation Today’s Readings
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Announcements more panorama slots available now
Announcements Questions on the project? New turn-in info online
Coupled Horn-Schunck and Lukas-Kanade for image processing
Announcements more panorama slots available now
Optical flow and keypoint tracking
Presentation transcript:

Feature Reconstruction Using Lucas-Kanade Feature Tracking and Tomasi-Kanade Factorization EE7740 Project I Dr. Gunturk

ABSTRACT Recovering 3-D structure from motion in noisy 2-D images is a problem addressed by many vision system researchers. By consistently tracking feature points of interest across multiple images using a methodology first described by Lucas- Kanade, a 3-D shape of the scene can be reconstructed using these features points using the factorization method developed by Tomasi-Kanade.

The image flow, or velocity field, in the image plane due to object/camera motion can be computed using feature matching. Velocity Flow Image I Image J x x d x + d

Total error E is the weighted sum-squared difference

Approximate I(x-d) using the Taylor series expansion A good match occurs when E is small, so we need to find a displacement d that minimizes E. This can be achieved by differentiating E with respect to d, setting it equal to zero, and solving for d. We can approximate the value of I(x-d) using the Taylor series expansion:

Approximate (cont.) first order term approx is sufficient for the calculations. Gradient of the intensity I is can represent the shifted intensity as sum-squared difference can now be represented as

Approximate (cont.) Taking the partial differentials with respect to x,y: equivalently

Approximate (cont.) Setting differential to 0 -> This can be represented in matrix form as Zd = e, where

corner detecting Harris filter “cornerness” function uses these 2 eigenvalues to give a quantitative measure of the corner and edge qualities.

Lucas-Kanade assumptions Z is invertible, that the two eigenvalues are large enough to be discernable from noise, and that the ratio of the two eigenvalues is well-behaved (larger/smaller is not too large). This is normally not the case.

desirable parameters for a tracker Accuracy can be related to the local sub-pixel resolution, in which a smaller integration window is desirable in order not to “smooth out” the details in the image. Robustness pertains to the sensitivity of the tracker to changes in lighting, size of image motion, etc. To handle larger motions, it is intuitive that a larger integration window would work better. One solution to this problem is a pyramidal Lucas-Kanade algorithm.

pyramidal Lucas-Kanade algorithm Using a Gaussian pyramid requires estimating the velocity at each pixel by solving Lucas-Kanade equations, using bilinear interpolation to warp the image so we keep all computation at a subpixel accuracy level, and then upsampling, continuing doing this same process for each layer of the pyramid all the way to the highest resolution (original image).

image I t-1 image I Gaussian pyramid of image I t-1 Gaussian pyramid of image I image I image I t-1 u=10 pixels u=5 pixels u=2.5 pixels u=1.25 pixels Coarse-to-fine optical flow estimation

image I image J Gaussian pyramid of image I t-1 Gaussian pyramid of image I image I image I t-1 Coarse-to-fine optical flow estimation run iterative L-K warp & upsample......

pseudo-code Goal: Let u be a point on image I. Find its corresponding location v on image J Build pyramid representations of I and J: {I L }L=0,…,L m and {J L }L=0,…,L m Initialization of pyramidal guess:

for L = L m down to 0 with step of -1 Location of point u on image I L : u L = [px py] T = u/2 L Derivative of I L with respect to x: I x (x, y) = I L (x + 1, y) - I L (x – 1, y) 2 Derivative of I L with respect to x: I x (x, y) = I L (x + 1, y) - I L (x – 1, y) 2 Spatial gradient matrix: Initialization of iterative L-K: for k = 1 to K with step of 1 (or until Image difference: Image mismatch vector: Optical flow (Lucas-Kanade): Guess for next iteration: end of for-loop on k < accuracy threshold)

Final optical flow at level L: Guess for next level L - 1: end of for-loop on L Final optical flow vector: d = g 0 + d 0 Location of point on J: v = u + d Solution: The corresponding point is at location v on image J

Initial Feature Points Methodology used to select the initial feature points on image I is as follows: Compute the G matrix and its minimum eigenvalue m at every pixel in image I. Determine the maximum max of all the minimum eigenvalues over the whole image. Retain the image pixels that have a m value that is 5%-10% of max. From those pixels keep the local max pixels (i.e. pixels are kept if its m value is larger than any other pixel in its 3x3 neighborhood). Keep the subset of those pixels so that the minimum distance between any pair of pixels is larger than a given threshold distance (typically 5 or 10 pixels).

Orthographic Case Trajectories of image coordinates {u fp,v fp } | f=1…F, p=1...P Input: registered measurement matrix Ŵ

The rank theorem place origin of the world coordinate at the centroid of the P points. Unit vectors i f,j f point along the direction X,Y of the image respectively

The rank theorem The projection (u fp,v fp ) i.e. the image feature point of point s p =(x p,y p,z p ) on to frame f  t f : the vector from world origin to the origin of image frame f Note: the origin is placed at the centroid of the object points, and since the origin of the world coordinates Is placed at the centroid of object points

The rank theorem For the registered horizontal image projection we have To summerize

The rank theorem The registered measurement matrix can be expressed in a matrix form: represents the camera rotation is the shape matrix

The rank theorem Since R is 2Fx3, S is 3xP, Rank theoremRank theorem: without noise, the registered measurement matrix is at most rank 3.

The registered measurement matrix Ŵ will be at most of rank three without noise. When noise corrupts the images, however, Ŵ will not be exactly of rank 3. The rank theorem can be extended to the case of noisy measurements in a well- defined manner, however, using approximate rank.

Approximate rank Ŵ can be decomposed into three matrix –Ŵ=O1∑O2, O1 and O2 are unitary matrix We have Ideally, ∑’ should contains all the singular value of Ŵ, O 1 ’’ ∑’’O 2 ’’ must be entirely to noise.

Rank theorem for Noisy Measurement All the shape and rotation information in W is contain in three greatest singular values, together with the corresponding left and right eigenvector.

Ř and Š same size as the desired rotation and shape matrices R and S decomposition is not unique (ŘQ)(Q-1Š) = Ř(QQ-1)Š = ŘŠ = Ŵ Since that column space is 3-D because of the rank theorem, R and Ř are different bases for the same space -> linear transformation between them Ř is a linear transformation of the true rotation matrix R Š is a linear transformation of the true rotation matrix S.

The metric constraints There exist a 3X3 matrix Q, –R= ŘQ, S=Q -1 Š To find Q: R is the rows of true rotation matrix. These metrix constraints yield the over- constrained quadratic system This is a simple nonlinear data fitting problem.

Experimental Results The 430 features selected by the automatic detection method Tomasi-Kanade.

Experimental Results 388 features selected by the automatic detection method Bishop 288 features tracked across 10 images by the automatic detection method Bishop Reconstructed image Bishop

Conclusions The pyramidal Lucas-Kanade tracker worked quite well on the images I submitted to it. For larger motions I would like to implement the Shi- Tomasi improvements I read about concerning an automatic scheme for rejecting spurious features in [7], but time constraints have not allowed for me to implement yet. The Tomasi-Kanade factorization method proved to be a robust solution for generating 3-D coordinates of feature points of rigid objects using the points tracked by the pyramidal Lucas- Kanade tracker.

References [1] "Pyramidal Implementation of the Lucas Kanade Feature Tracker Description of the algorithm", Jean-Yves Bouguet, Intel Corporation, Microprocessor Research Labs, [2] "A combined corner and edge detector", Chris Harris and Mike Stephens, Proceedings Fourth Alvey Vision Conference, Manchester, pp , [3] “Good Features to Track”, Jianbo Shi and Carlo Tomasi, IEEE Conference on Computer Vision and Pattern Recognition (CVPR94), Seattle, June 1994 [4] “Shape and motion from image streams under orthography: a factorization method.” Carlo Tomasi and Takeo Kanade, International Journal of Computer Vision, 9(2): , November [5] [6] “Linear and Incremental Acquisition of Invariant Shape Models from Image Sequences”, Daphna Weinshall and Carlo Tomasi, Proceedings: IEEE fourth International Conference of Computer Vision, pp , Berlin, May [7] “Improving Feature Tracking with Robust Statistics”, A. Fusiello, E. Trucco, T. Tommasini, V. Roberto, Pattern Analysis & Applications (1999)2:312–320, Ó 1999 Springer-Verlag London Limited