Image Based Rendering And Modeling Techniques And Their Applications Jiao-ying Shi State Key laboratory of Computer Aided Design and Graphics Zhejiang.

Slides:



Advertisements
Similar presentations
The fundamental matrix F
Advertisements

3D reconstruction.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Computer vision: models, learning and inference
Two-View Geometry CS Sastry and Yang
Camera calibration and epipolar geometry
Structure from motion.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
CSCE 641 Computer Graphics: Image-based Modeling Jinxiang Chai.
Copyright  Philipp Slusallek IBR: View Interpolation Philipp Slusallek.
Siggraph’2000, July 27, 2000 Jin-Xiang Chai Xin Tong Shing-Chow Chan Heung-Yeung Shum Microsoft Research, China Plenoptic Sampling SIGGRAPH’2000.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Projected image of a cube. Classical Calibration.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Lec 21: Fundamental Matrix
CSE473/573 – Stereo Correspondence
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Today: Calibration What are the camera parameters?
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #15.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Automatic Camera Calibration
Camera Calibration & Stereo Reconstruction Jinxiang Chai.
1 Intelligent Robotics Research Centre (IRRC) Department of Electrical and Computer Systems Engineering Monash University, Australia Visual Perception.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Image-Based Rendering. 3D Scene = Shape + Shading Source: Leonard mcMillan, UNC-CH.
Camera Geometry and Calibration Thanks to Martial Hebert.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Image Based Rendering(IBR) Jiao-ying Shi State Key laboratory of Computer Aided Design and Graphics Zhejiang University, Hangzhou, China
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
视觉的三维运动理解 刘允才 上海交通大学 2002 年 11 月 16 日 Understanding 3D Motion from Images Yuncai Liu Shanghai Jiao Tong University November 16, 2002.
3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: , Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
CSCE 643 Computer Vision: Structure from Motion
Geometric Camera Models
Binocular Stereo #1. Topics 1. Principle 2. binocular stereo basic equation 3. epipolar line 4. features and strategies for matching.
Plenoptic Modeling: An Image-Based Rendering System Leonard McMillan & Gary Bishop SIGGRAPH 1995 presented by Dave Edwards 10/12/2000.
© 2005 Martin Bujňák, Martin Bujňák Supervisor : RNDr.
Lecture 03 15/11/2011 Shai Avidan הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Affine Structure from Motion
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Single-view geometry Odilon Redon, Cyclops, 1914.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
3D Imaging Motion.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
112/5/ :54 Graphics II Image Based Rendering Session 11.
EECS 274 Computer Vision Affine Structure from Motion.
1 Chapter 2: Geometric Camera Models Objective: Formulate the geometrical relationships between image and scene measurements Scene: a 3-D function, g(x,y,z)
Feature Matching. Feature Space Outlier Rejection.
Computer vision: models, learning and inference M Ahad Multiple Cameras
CSCE 641 Computer Graphics: Image-based Rendering (cont.) Jinxiang Chai.
3D Reconstruction Using Image Sequence
3D Sensing 3D Shape from X Perspective Geometry Camera Model Camera Calibration General Stereo Triangulation 3D Reconstruction.
Geometry Reconstruction March 22, Fundamental Matrix An important problem: Determine the epipolar geometry. That is, the correspondence between.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
Introduction To IBR Ying Wu. View Morphing Seitz & Dyer SIGGRAPH’96 Synthesize images in transition of two views based on two images No 3D shape is required.
Computer vision: geometric models Md. Atiqur Rahman Ahad Based on: Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Computer vision: models, learning and inference
Advanced Computer Graphics
Image-Based Rendering
© 2005 University of Wisconsin
Epipolar geometry.
Common Classification Tasks
Single-view geometry Odilon Redon, Cyclops, 1914.
Presentation transcript:

Image Based Rendering And Modeling Techniques And Their Applications Jiao-ying Shi State Key laboratory of Computer Aided Design and Graphics Zhejiang University, Hangzhou, China

Image Based Rendering (IBR) PART I

Traditional Computer Graphics Use geometry and lighting model to simulate the imaging process and generate realistic scene –No guarantees for the rightness of the models –A lot of computation time needed

Use of Images in Computer Graphics Texture mapping Environment map How about more images?

Image Based Rendering IBR: To Synthesize a new scene with novel view point based on given images

A Framework of Image Based Rendering Real Scene Sampling System Data Storage System Data representation System Rendering System Synthesized view

The Key Part of IBR The data representation system is the key part of IBR, It determines the other three subsystems. - A taxonomy based on the data representation system

A Taxonomy of IBR The Geometry based data representation The Image based data representation The plenoptic function based data representation

The Geometry Based Data Representation Geometry elements used as data representation in IBR: –polyhedra(Debevec, et. al 1996) –layers (Baker, Szeliski and Anandan 1998) –points(Shade et al. 1998) Similar to traditional Computer Graphics, except the geometry model comes from images

Image Based Data Representation data are treated as a series of images with correspondence relations “ optical flow ” “ morphing map ” forward/ reverse mapping;morphing Examples:. View interpolation (Chen and William,1993).View Morphing(Seitz and Dyer 1998)

Plenoptic Function Based Data Representation Plenoptic function (Adelson and Bergen,1991)

Representative IBR Methods Based on Plenoptic Functions Plenoptic Modeling: 5D Light field/Lumigraph: 4D Concentric Mosaics : 3D Panorama: 2D [L. McMillan 95] [M. Levoy 96, S. J. Gortler 96] [H. Y. Shum 99] [S. E. Chen 95, R. Szeliski 97]

Conclusion The progress of IBR technique is also the progress of new data representation method. We treat an image: –as texture in geometry  texture mapping –as images with correspondence relation  view interpolation /morphing –as light beams  light field –as slit image  concentric mosaics...

Demo of IBR (1) Tour in Dunhuang Art Cave

Demo of IBR (2) Tour in Lingyin Temple in Hangzhou

Image Based Modeling (IBM) PART II

The Common Methods Used for Modeling objects Using Geometry Modeling Software Packages, such as 3D MAX , Mayer , SoftImage and AutoCAD etc. to create wireframe models, surface models or volume models. Using 3D Laser Scanners Using Image-based Modeling techniques to create geometry models or appearance models of objects.

A Taxonomy of IBM IBM methods using active cues The active cues are refered to the artificially generated silhouette, which are projected onto the surface of the modeled objects. IBM methods using passive cues The passive cues are refered to the implicit characteristics of the modeled objects, such as geometry features and textures of the objects.

IBM Methods Using Active cues

IBM Methods Using Passive Cues Based on known geometry Based on visual hull Based on light field Based on stereo vision

IBM Methods Using Passive Cues Based on known geometry[Debevec96]

IBM Methods Using Passive Cues Based on visual hull [Wojciech Matusik 2001]

IBM Methods Using Passive Cues Based on Light Field[Marc Pollefeys et. al.]

IBM Methods Using Passive Cues Based on stereo pairs: the goal is to automatically extract a realistic 3D model by freely moving a camera around an object.  Neither the camera motion nor the camera settings have to be known.  The obtained 3D model is a scaled version of the original object.  The surface appearance is obtained from the image sequence as well.

IBM Methods Using Passive Cues How can we get 3D information (depth) from a 2D image? It is impossible to get depth information of the object from a 2D image.

IBM Methods Using Passive Cues It is possible to recover the depth information of the object by using two or more images through triangulation. Reconstruction of 3D point through triangulation

IBM Methods Using Passive Cues But the following information are needed to know for this purpose: Corresponding image points Relative pose of the camera for different views(so called camera extrinsic parameters, i.e. the position and orientation of the camera) Relation between the image points and the corresponding line of sight. It is defined by the camera model, which usually is a pinhole model and cameras intrinsic parameters and extrinsic parameters.

IBM Methods Using Passive Cues Flowchart of the IBM technique based on stereo pairs

Reconstruction of Architectural Models Based on Image Sequence

Camera Calibration Using Corner Structure and Parallel Structure in Real Scene Consisted of at least 4 line segments. 3 of them are perpendicular each other. The forth segment is parallel to one of previous 3 segments. Corner Structure Parallel Structure Consisted of at least 4 line segments. 2 of them are perpendicular each other. Other 2 segments are parallel to previous 2 segments separately.

Calculating focal length based on single image is the directional vector of segment ca is the normal vector of the projective plane is the directional vector of segment AB because We can get (*) This is a quartic equation relative to f, from which we can get one plus real solution of the focal length. Projection of the corner structure Then the directional vectors of OX and OY can be written as an expression relative to the focal length f

Calculating rotation matrix and translation vector based on single image related to SCF From the projection equation of pinhole camera model, we can get: Rotation matrix relative to SCF is: The direction of the translation vector relative to CCF is Projection of the corner structure SCF: structure coordinate frame CCF: camera coordinate frame Translation vector can be deduced:

Optimizing the camera ’ s parameters: f, r and t We choose the distance DIST between the original image point and the reprojected image point as the measurement function for error. Once the initial values of camera parameters were derived, we can optimize these parameters by using other line segments, which are parallel to the segments of the corner structure in two images.

Procedures for optimization  Rotation matrices  Translation vector  Structure condition  Objective function of optimization: To ensure the orthogonality of rotation matrices To ensure the length of the two transla- tion vectors will not be changed in optimization. tlen1 and tlen2 are the modules of the two translation vectors, which can be calculated from the initial solutions. To ensure directions of the segments will not be changed in optimization  Constrained conditions:

Experimental Results Two pictures of kiosk at the campus of Zhejiang University taken by hand-hold camera Camera positions and top-down views of feature points

Error Statistics: Choose the distance between feature points to its epipolar line as the measurement of error. In the ideal situation this distance should be equal to zero. Image Err before optimization (pixels) Err after optimization (pixels) Left Right Metric(Pixel)

Architecture Reconstruction Based on Single Image

Reconstruction of plane Assign base point O and determine its coordinates Assign base plane which contains the base point O as one of its corners. Determine the normal vector of the base plane. Determine unknown neighbor plane of the base plane: choose one point on the common edge of the neighbor planes as the base point of the unknown plane. Determine normal vector of the unknown plane.

Decomposition of sweeping surface into plane patches

Wireframe model

Texture mapping model

Strategy of model merging for multiple scenes Scene structure and 3 camera positions Model merging diagram for multiple scenes

Example of model merging Image sequence Reconstructed model for each image

Main challenges for model merging procedure How to transform all the models in different local coordinates to one world coordinate frame? Model integration: scaling, vertex match? Deletion of overlapped plane?

Reconstructed Tea Box with Texture Mapped

Experiment 1: image sequence

Experiment 1: reconstructed model

Experiment 2: image sequence

Experiment 2: reconstructed model

Thanks a lot