Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.

Slides:



Advertisements
Similar presentations
Motion Estimation I What affects the induced image motion? Camera motion Object motion Scene structure.
Advertisements

MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Two-View Geometry CS Sastry and Yang
Jan-Michael Frahm, Enrique Dunn Spring 2012
Two-view geometry.
Lecture 8: Stereo.
Computer Vision Optical Flow
Camera calibration and epipolar geometry
Structure from motion.
Motion Estimation I What affects the induced image motion? Camera motion Object motion Scene structure.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Multiple-view Reconstruction from Points and Lines
CSc83029 – 3-D Computer Vision/ Ioannis Stamos 3-D Computational Vision CSc Optical Flow & Motion The Factorization Method.
Feature tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on good features.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.
COMP 290 Computer Vision - Spring Motion II - Estimation of Motion field / 3-D construction from motion Yongjik Kim.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
KLT tracker & triangulation Class 6 Read Shi and Tomasi’s paper on good features to track
Optical Flow Digital Photography CSE558, Spring 2003 Richard Szeliski (notes cribbed from P. Anandan)
3D Computer Vision and Video Computing 3D Vision Topic 8 of Part 2 Visual Motion (II) CSC I6716 Spring 2004 Zhigang Zhu, NAC 8/203A
Structure Computation. How to compute the position of a point in 3- space given its image in two views and the camera matrices of those two views Use.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Multi-view geometry. Multi-view geometry problems Structure: Given projections of the same 3D point in two or more images, compute the 3D coordinates.
Automatic Camera Calibration
Computer vision: models, learning and inference
Image Stitching Ali Farhadi CSE 455
Image alignment.
CSC 589 Lecture 22 Image Alignment and least square methods Bei Xiao American University April 13.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.
The Brightness Constraint
The Measurement of Visual Motion P. Anandan Microsoft Research.
Image stitching Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz, Matthew Brown and Vaclav Hlavac.
Multiview Geometry and Stereopsis. Inputs: two images of a scene (taken from 2 viewpoints). Output: Depth map. Inputs: multiple images of a scene. Output:
Uses of Motion 3D shape reconstruction Segment objects based on motion cues Recognize events and activities Improve video quality Track objects Correct.
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
3D Imaging Motion.
Two-view geometry. Epipolar Plane – plane containing baseline (1D family) Epipoles = intersections of baseline with image planes = projections of the.
Feature Matching. Feature Space Outlier Rejection.
Motion Estimation I What affects the induced image motion?
Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in.
MASKS © 2004 Invitation to 3D vision Uncalibrated Camera Chapter 6 Reconstruction from Two Uncalibrated Views Modified by L A Rønningen Oct 2008.
Uncalibrated reconstruction Calibration with a rig Uncalibrated epipolar geometry Ambiguities in image formation Stratified reconstruction Autocalibration.
776 Computer Vision Jan-Michael Frahm Spring 2012.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Lecture 16: Image alignment
Motion and Optical Flow
René Vidal and Xiaodong Fan Center for Imaging Science
The Brightness Constraint
3D Motion Estimation.
Epipolar geometry.
Image Stitching Slides from Rick Szeliski, Steve Seitz, Derek Hoiem, Ira Kemelmacher, Ali Farhadi.
3D Photography: Epipolar geometry
Structure from motion Input: Output: (Tomasi and Kanade)
The Brightness Constraint
The Brightness Constraint
Uncalibrated Geometry & Stratification
Filtering Things to take away from this lecture An image as a function
Computational Photography
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Linearizing (assuming small (u,v)): Brightness Constancy Equation: The Brightness Constraint Where:),(),(yxJyxII t  Each pixel provides 1 equation in 2 unknowns (u,v). Insufficient info. Another constraint: Global Motion Model Constraint

Camera induced motion + = Independent motions 3D Camera motion + 3D Scene structure + Independent motions The 2D/3D Dichotomy Image motion = 2D techniques 3D techniques Singularities in “2D scenes” Do not model “3D scenes”  Requires prior model selection

The only part with 3D depth information The 2D/3D Dichotomy When cannot recover any 3D info? Planar scene: In the uncalibrated case (unknown calibration matrix K)  Cannot recover 3D rotation or Plane parameters either (because cannot tell the difference between H and KR)

Global Motion Models 2D Models: 2D Similarity 2D Affine Homography (2D projective transformation) 3D Models: 3D Rotation + 3D Translation + Depth Essential/Fundamental Matrix Plane+Parallax  Relevant when camera is translating, scene is near, and non-planar.  Relevant for: *Airborne video (distant scene) * Remote Surveillance (distant scene) * Camera on tripod (pure Zoom/Rotation) * 2D models always provide dense correspondences. * 2D Models are easier to estimate than 3D models (much fewer unknowns  numerically more stable).

Example: Affine Motion Substituting into the B.C. Equation: Each pixel provides 1 linear constraint in 6 global unknowns (minimum 6 pixels necessary) Least Square Minimization (over all pixels): Every pixel contributes  Confidence-weighted regression

Example: Affine Motion Differentiating w.r.t. a 1, …, a 6 and equating to zero  6 linear equations in 6 unknowns: Summation is over all the pixels in the image!

image I image J JwJw warp refine + Pyramid of image JPyramid of image I image I image J Coarse-to-Fine Estimation u=10 pixels u=5 pixels u=2.5 pixels u=1.25 pixels ==> small u and v... Parameter propagation:

Other 2D Motion Models 2D Projective – planar motion (Homography H)

Panoramic Mosaic Image Original video clip Generated Mosaic image Alignment accuracy (between a pair of frames): error < 0.1 pixel

Original Outliers Original Synthesized Video Removal

ORIGINAL ENHANCED Video Enhancement

Direct Methods: Methods for motion and/or shape estimation, which recover the unknown parameters directly from image intensities.  Error measure based on dense image quantities (Confidence-weighted regression; Exploits all available information) Feature-based Methods: Methods for motion and/or shape estimation based on feature matches (e.g., SIFT, HOG).  Error measure based on sparse distinct features (Features matches + RANSAC + Parameter estimation)

Image gradients The descriptor (4x4 array of 8-bin histograms) –Compute gradient orientation histograms of several small windows (128 values for each point) –Normalize the descriptor to make it invariant to intensity change –To add Scale & Rotation invariance: Determine local scale (by maximizing DoG in scale and in space), local orientation as the dominant gradient direction. Example: The SIFT Descriptor D.Lowe. “Distinctive Image Features from Scale-Invariant Keypoints”. IJCV 2004 Compute descriptors in each image Find descriptors matches across images  Estimate transformation between the pair of images. In case of multiple motions: Use RANSAC (Random Sampling and Consensus) to compute Affine-transformation / Homography / Essential-Matrix / etc.

Benefits of Direct Methods High subpixel accuracy. Simultaneously estimate matches + transformation  Do not need distinct features for image alignment: Strong locking property.

Limitations of Direct Methods Limited search range (up to ~10% of the image size). Brightness constancy assumption.

DEMO: Video Indexing and Editing Exercise 4: Image alignment (will be posted in a few days) Keep reference image the same (i.e., unwarp target image)  Estimate derivatives only once per pyramid level. Avoid repeated warping of the target image  Accumulate translations and unwarp target image once.

The 2D/3D Dichotomy Image motion = Camera induced motion = + Independent motions = Camera motion + Scene structure + Independent motions 2D techniques 3D techniques Singularities in “2D scenes” Do not model “3D scenes” Source of dichotomy: Camera-centric models (R,T,Z)

The Plane+Parallax Decomposition Original SequencePlane-Stabilized Sequence The residual parallax lies on a radial (epipolar) field: epipole Move from CAMERA-centric to a SCENE-centric model

Benefits of the P+P Decomposition Eliminates effects of rotation Eliminates changes in camera calibration parameters / zoom Camera parameters: Need to estimate only the epipole. (i.e., 2 unknowns) Image displacements: Constrained to lie on radial lines (i.e., reduces to a 1D search problem)  A result of aligning an existing structure in the image. 1. Reduces the search space:

Remove global component which dilutes information ! Translation or pure rotation ??? Benefits of the P+P Decomposition 2. Scene-Centered Representation: Focus on relevant portion of info

Benefits of the P+P Decomposition 2. Scene-Centered Representation: Shape = Fluctuations relative to a planar surface in the scene STAB_RUG SEQ

- fewer bits, progressive encoding Benefits of the P+P Decomposition 2. Scene-Centered Representation: Shape = Fluctuations relative to a planar surface in the scene Height vs. Depth (e.g., obstacle avoidance) A compact representation global (100) component local [-3..+3] component total distance [ ] camera center scene Appropriate units for shape

Start with 2D estimation (homography). 3D info builds on top of 2D info. 3. Stratified 2D-3D Representation: Avoids a-priori model selection. Benefits of the P+P Decomposition

Original sequencePlane-aligned sequenceRecovered shape Dense 3D Reconstruction (Plane+Parallax)

Original sequence Plane-aligned sequence Recovered shape

Original sequence Plane-aligned sequence Recovered shape Dense 3D Reconstruction (Plane+Parallax)

Brightness Constancy constraint P+P Correspondence Estimation The intersection of the two line constraints uniquely defines the displacement. 1. Eliminating Aperture Problem Epipolar line epipole p

other epipolar line Epipolar line Multi-Frame vs. 2-Frame Estimation The two line constraints are parallel ==> do NOT intersect 1. Eliminating Aperture Problem p another epipole Brightness Constancy constraint The other epipole resolves the ambiguity !