# Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.

## Presentation on theme: "Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002."— Presentation transcript:

Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002

Problem Statement How can one extract information from a sequence of images without camera calibration? Sequence of images World Model

Panoramic Mosaicing Image Sequence

Projective Depth Recovery d a1 d b1 a b 12 3 Image Sequence

Ambiguity What does the 3-D structure look like?

Ambiguity What does the 3-D structure look like? d1d1 d2d2 d3d3 d4d4

Ambiguity What does the 3-D structure look like? d1d1 d2d2 d3d3 d4d4 The projective depth is defined up to a projective transform.

Ambiguity What does the 3-D structure look like?

Outline Image transformations. Direct methods for image registration. Mosaic construction. Projective depth recovery.

2D Transformations Planar Scene Image Plane How does the square look when we move the camera?

Types of Transformations Rigid Rigid+Scaling Affine Projective 2D Examples

Mathematically The 2-D planar transformation p’ of point p is p’p’ p M 2D =

3-D Rigid Transformations 3-D Scene Rotation + translation u u’u’ How do u and u’ relate?

Mathematically Calculate the world coordinate p from an image coordinate u: Calculate image coordinate u’ from p: Viewing matrix M

Panoramas What if the optical center of the camera does not move? The images are related by a homography.

Direct Image Registration For a pair of images, I and I’, minimize the intensity discrepancy

Nonlinear Iterative Minimization At the minimum of E, Jacobian

Nonlinear Iterative Minimization At the minimum of E, Given some estimate of find s.t. Update :, where Hessian

Levenberg-Marquardt The Hessian is hard to calculate. Approximate it as Levenberg-Marquardt finds locally optimal solutions.

Hierarchical Matching Pyramid of image I Pyramid of image I’ Image I’ Image I Subsample Calculate M a M a M b M b M c M c M

Mosaic Construction How do we stitch together the registered images? One approach: Choose one frame’s coordinate system. Warp all frames into that coordinate system. Blend together overlapping sections by averaging.

Mosaic Constuction Example

Environment Maps Color each face a different color. Unroll into a 2D texture map

Environment Maps Expand each face.

Environment Maps For each face, determine the mapping to world coordinates s.t. u1u1 u2u2 u3u3 u4u4 p1p1 p2p2 p3p3 u i = M F p i

Environment Maps Warp each image to the coordinate system of each face. For each face, form a blended image. Paint the blended image faces into the 2D texture map. This method can be performed for arbitrary surfaces, including a tesselated sphere.

Cubical Environment Map

Tesselated Sphere Results

Blending How do we choose pixel values in the mosaic where the images overlap? Superimposing method: Inconsistencies at frame boundaries

Blending How do we choose pixel values in the mosaic where the images overlap? Weighted averaging:

Blending How do we choose pixel values in the mosaic where the images overlap? Weighted averaging: Overlapping images are averaged, weighted by distance from the center.

Blending How do we choose pixel values in the mosaic where the images overlap? Multi-resolution blending: Overlapping images are averaged, weighted by proximity to desired zoom.

Projective Depth Recovery Earlier, we saw that if two cameras are related by a 3-D rigid transformation,

Projective Depth Recovery Earlier, we saw that if two cameras are related by a 3-D rigid transformation, Image coordinate in I’ Image coordinate in I Homography Viewing matrix for I’ Projective depth Translation from O to O’

Projective Depth Recovery Earlier, we saw that if two cameras are related by a 3-D rigid transformation, Image coordinate in I’ Image coordinate in I Homography Projective depth Parallax motion

Projective Depth Recovery Earlier, we saw that if two cameras are related by a 3-D rigid transformation, Image coordinate in I’ Image coordinate in I Homography Projective depth Parallax motion

Algorithmic Idea Choose a base frame I 0 to recover projective depth in. Find M j,, and d i to minimize using nonlinear iterative minimization.

Ambiguity

Number of Parameters How many parameters must we estimate? (8 + 3) n + p, where n is the number of images and p is the number of pixels. p is large, so the depth map is represented using a tensor-product spline.

Splines Pixel ( i,j ) Spline control vertex Depth Map Spline Depth i j Let the depths at the control vertices vary. Find the depths at all points by interpolation.

Local Minima The high dimensionality of the search space increases the chance of finding a nonglobal optimum. One solution is to initialize the dense algorithm with the results of a feature- based algorithm.

Feature-Based Algorithm Detect features, for example corners, in each frame. Find between-frame feature correspondences.

Feature-Based Algorithm Now we have the locations of each feature i in each frame j, v ij. Find the transformation ( M j,, d i ) to minimize through nonlinear iterative minimization.

Feature-Based Algorithm Now we have the locations of each feature i in each frame j, v ij. Find the transformation ( M j,, d i ) to minimize through nonlinear iterative minimization. Inverse variance weight Location of feature i in base frame

Algorithm Initialization Simple approach: initialize Faster approach: Fixand, then solve for M j. Estimate and

View Interpolation Input image Virtual camera Novel image Input image

View Interpolation We can approximate the Euclidean depth map from the projective depth map. From the Euclidean depth map, we can synthesize novel views of a scene.

View Interpolation Results Image sequence taken by moving camera up and down Dense depth map Novel interpolated view

View Interpolation Results Input image sequence Dense depth map Novel interpolated view

Summary Direct methods can be used to register image frames if between-frame motion is small without camera calibration. Once the frames are registered, mosaics can be constructed if the camera doesn’t translate. Projective depth can be recovered from a sequence of images if the camera translates. Euclidean depth can be estimated from projective depth. Novel views can be synthesized from Euclidean depth.

References R. Szeliski, P. Anandan, K. Toyama, and L. Kanade,. Notes from Vision for Graphics, CSE 590SS, University of Washington, Winter 2001. R. Szeliski and S. B. Kang. Direct methods for visual scene reconstruction. In IEEE Workshop on Repres. of Visual Scenes, pages 26-33, June 1995. R. Szeliski. Image Mosaicing for Tele-Reality Applications. Technical Report 94/2, Digital Equipment Corporation, Cambridge Research Lab, June 1994. R. Szeliski and J. Coughlan. Spline-Based Image Registration. Technical Report 94/1, Digital Equipment Corporation, Cambridge Research Lab, April 1994. R. Szeliski and S. B. Kang. Recovering 3D shape and motion from image streams using non-linear least squares. Journal of Visual Communication and Image Representation, Vol. 5(1), pages 10-28, 1994.

References R. Szeliski and H. Shum. Creating full view panoramic image mosaics and environment maps. In Proc. of SIGGRAPH, pages 251-258, 1997. H. Shum and R. Szeliski. Construction of Panoramic Image Mosaics with Global and Local Alignment. In ICCV Vol. 36(2), pages 101-130, 2000. D. Capel and A. Zisserman. Automated Mosaicing with Super- Resolution Zoom. In Proc. CVPR, pages 885-891, Jun 1998. M. Irani and P. Anandan. All about direct methods. In W. Triggs, A. Zisserman, and R. Szeliski, editors, Vision Algorithms: Theory and practice. Springer-Verlag, 1999. R. Szeliski. Video Mosaics for Virtual Environments. In IEEE Computer Graphics, Volume 16(2), pages 22-30, March 1996.

Download ppt "Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002."

Similar presentations