CSE473/573 – Stereo and Multiple View Geometry

Slides:



Advertisements
Similar presentations
Computer Vision, Robert Pless
Advertisements

The fundamental matrix F
Lecture 11: Two-view geometry
3D reconstruction.
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
Gratuitous Picture US Naval Artillery Rangefinder from World War I (1918)!!
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Lecture 8: Stereo.
Dr. Hassan Foroosh Dept. of Computer Science UCF
Camera calibration and epipolar geometry
Structure from motion.
Last Time Pinhole camera model, projection
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
Lecture 21: Multiple-view geometry and structure from motion
Multi-view stereo Many slides adapted from S. Seitz.
CSCE 641 Computer Graphics: Image-based Modeling Jinxiang Chai.
Stereo and Structure from Motion
Lecture 20: Two-view geometry CS6670: Computer Vision Noah Snavely.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Lec 21: Fundamental Matrix
CSE473/573 – Stereo Correspondence
Review: Binocular stereo If necessary, rectify the two stereo images to transform epipolar lines into scanlines For each pixel x in the first image Find.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Instance-level recognition I. - Camera geometry and image alignment Josef Sivic INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Automatic Camera Calibration
Epipolar Geometry and Stereo Vision Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/05/15 Many slides adapted from Lana Lazebnik,
Computer vision: models, learning and inference
Camera Calibration & Stereo Reconstruction Jinxiang Chai.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Structure from images. Calibration Review: Pinhole Camera.
Final Exam Review CS485/685 Computer Vision Prof. Bebis.
Camera Geometry and Calibration Thanks to Martial Hebert.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
Recap from Monday Image Warping – Coordinate transforms – Linear transforms expressed in matrix form – Inverse transforms useful when synthesizing images.
CSCE 643 Computer Vision: Structure from Motion
Geometric Camera Models
Stereo Many slides adapted from Steve Seitz.
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image image 1image.
Computer Vision, Robert Pless
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Announcements Project 3 due Thursday by 11:59pm Demos on Friday; signup on CMS Prelim to be distributed in class Friday, due Wednesday by the beginning.
Peripheral drift illusion. Multiple views Hartley and Zisserman Lowe stereo vision structure from motion optical flow.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
stereo Outline : Remind class of 3d geometry Introduction
1 Chapter 2: Geometric Camera Models Objective: Formulate the geometrical relationships between image and scene measurements Scene: a 3-D function, g(x,y,z)
Feature Matching. Feature Space Outlier Rejection.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Solving for Stereo Correspondence Many slides drawn from Lana Lazebnik, UIUC.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Computer vision: models, learning and inference M Ahad Multiple Cameras
776 Computer Vision Jan-Michael Frahm & Enrique Dunn Spring 2013.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
CSE 185 Introduction to Computer Vision Stereo 2.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
제 5 장 스테레오.
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
What have we learned so far?
3D reconstruction class 11
Multiple View Geometry for Robotics
Uncalibrated Geometry & Stratification
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Stereo vision Many slides adapted from Steve Seitz.
Presentation transcript:

CSE473/573 – Stereo and Multiple View Geometry Presented by Radhakrishna Dasari

Contents Stereo Practical Demo Camera Intrinsic and Extrinsic parameters Essential and Fundamental Matrix Multiple View Geometry Multi-View Applications

Stereo Vision Basics Stereo Correspondence – Epipolar Epipolar constraint Rectification Pixel matching Depth from Disparity C. Loop and Z. Zhang. Computing Rectifying Homographies for Stereo Vision. IEEE Conf. Computer Vision and Pattern Recognition, 1999.

Stereo Rectification Rectification is the process of transforming stereo images, such that the corresponding points have the same row coordinates in the two images. It is a useful procedure in stereo vision, as the 2-D stereo correspondence problem is reduced to a 1-D problem Let’s see the rectification pipeline when we have are two images of the same scene taken from a camera from different viewpoints

Stereo Input Images Superposing the two input images on each other and compositing

Matching Feature Points

Eliminating outliers using RANSAC We can impose geometric constraints while applying RANSAC for eliminating outliers

Estimate Fundamental Matrix using Matched Points fMatrix = estimateFundamentalMatrix( matchedPtsOut.Location, matchedPtsIn.Location);

Rectified Input Stereo Images

Depth From Disparity

Rectified Stereo Images as Input

Disparity map using Block Matching There are noisy patches and bad depth estimates, especially on the ceiling. These are caused when no strong image features appear inside of the pixel windows being compared. The matching process is subject to noise since each pixel chooses its disparity independently of all the other pixels.

Disparity map using Dynamic Programming – Simple Example For optimal path we use the underlying block matching metric as the cost function constrain the disparities to only change by a certain amount between adjacent pixels (Smoothness of disparity) Lets say +/- 3 values of the neighbors We assign a penalty for disparity disagreement between neighbors. Hence most of the noisy blocks will be eliminated. Good matches will be preserved as block-matching cost function will dominate the penalty assigned for disparity disagreement

Depth from Disparity and Back-Projection With a stereo depth map and knowledge of the intrinsic parameters (focal length, image center) of the camera, it is possible to back-project image pixels into 3D points Intrinsic Parameters of a camera are obtained using camera calibration techniques

Camera Intrinsic Parameters Camera Calibration Matrix ‘K’ – 3x3 Upper triangular Matrix Constitutes – Focal length of the camera ‘f’ , Principal Point (u0,v0), aspect ratio of the pixel ‘γ’ and the skew ‘s’ of the sensor pixel Intrinsic parameters can be estimated using camera calibration techniques Ideal image sensor Sensor pixel with skew

Camera Calibration with grid templates Camera Calibration Toolbox on Matlab

Intrinsic & Extrinsic Parameters The transformation of point ‘pw’ from world is related to the point on image plane ‘x’ through the Projection Matrix ‘P’ which constitutes intrinsic and extrinsic parameters Camera matrix – both intrinsic ‘K’ (focal length, principal point) and extrinsic parameters (Pose – ‘R’ rotation matrix and ‘t’ translation) Projection Matrix or Camera Matrix ‘P’ is of dimension ‘3x4’

Projection Matrix ‘P’ Special case of perspective projection – Orthographic Projection Also called “parallel projection”: (x, y, z) → (x, y) What’s the projection matrix? Image World

Projection Matrix ‘P’ In general, for a perspective projection Matrix ‘P’ maps image point ‘x’ into world co-ordinates ‘X’ as The Projection Matrix (3x4) can be decomposed into (3x4) (3x3) (3x4) (4x4) (4x4)

Pure Rotational Model of Camera - Homography α,β,γ are angle changes across roll, pitch and yaw

Homography Suppose we have two images of a scene captured from a rotating camera point ‘x1 ’ in Image1 is related to the world point ‘X’ by the equation x1 = KR1X which implies X = R1-1K-1 x1 as point ‘x2 ’ in Image2 is related to the world point ‘X’ by the equation x2 = KR2X = KR2R1-1K-1 * x1 Hence the points in both the images are related to each other by a transformation of Homography ‘H’ x2 = H x1 Where H = KR2R1-1K-1

Rotation of Camera along Pitch, Roll and Yaw If the camera is only rotating along these axes and there is zero translation, the captured images can be aligned with each other using Homography estimation The Homography Matrix ‘H’ (3x3)can be estimated by matching features between two images

Image Alignment Result - Rotation of Camera along Pitch Axis

Image Alignment Result- Rotation of Camera along Roll axis

Image Alignment Result- Rotation of Camera along Yaw axis

Fundamental and Essential Matrices the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. Fundamental and Essential Matrices Stereo Images have both rotation and translation of camera the fundamental matrix ‘F’ is a 3×3 matrix which relates corresponding points x and x1 in stereo images. It captures the essence of Epipolar constraint in the Stereo images. Essential Matrix Where K and K1 are the Intrinsic parameters of the cameras capturing x and x1 respectively http://en.wikipedia.org/wiki/Eight-point_algorithm

Stereo – Fundamental and Essential Matrices https://www.youtube.com/watch?v=DgGV3l82NTk

Beyond Two-View Stereo the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. Beyond Two-View Stereo Third View can be used for verification

Multi-View Video in Dynamic Scenes Reference link

Multiple-View Geometry the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. Multiple-View Geometry Generic problem formulation: given several images of the same object or scene, compute a representation of its 3D shape

Multiple-baseline Stereo the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. Multiple-baseline Stereo Pick a reference image, and slide the corresponding window along the corresponding epipolar lines of all other images using other images Remember? disparity Where B is baseline, f is focal length and Z is the depth This equation indicates that for the same depth the disparity is proportional to the baseline M. Okutomi and T. Kanade, “A Multiple-Baseline Stereo System,” IEEE Trans. on Pattern Analysis and Machine Intelligence, 15(4):353-363 (1993)

Feature Matching to Dense Stereo the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. Feature Matching to Dense Stereo 1. Extract features 2. Get a sparse set of initial matches 3. Iteratively expand matches to nearby locations Iteratively expand matches to nearby locations 4. Use visibility constraints to filter out false matches 5. Perform surface reconstruction

the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. View Synthesis Is it possible to synthesize views from the locations where the cameras are removed? i.e Can we synthesize view from a virtual camera

View Synthesis - Basics the fundamental matrix   is a 3×3 matrix which relates corresponding points in stereo images. View Synthesis - Basics Problem: Synthesize virtual view of the scene at the mid point of line joining Stereo camera centers. Given stereo images, find Stereo correspondence and disparity estimates between them.

View Synthesis - Basics Use one of the images and its disparity map to render a view at virtual camera location. By shifting pixels with half the disparity value

View Synthesis - Basics Use the information from other image to fill in the holes, by shifting the pixels by half the disparity

View Synthesis - Basics Putting both together, we have the intermediary view. We still have holes. Why??

View Synthesis – Problem of Holes

View Synthesis – Problem of Color Variation at boundaries

Slide Credits Rob Fergus, S Seitz, Lazebnik MATLAB Computer Vision Toolbox