Plenoptic Stitching: A Scalable Method for Reconstructing 3D Interactive Walkthroughs Daniel G. Aliaga Ingrid Carlbom

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

The fundamental matrix F
Lecture 11: Two-view geometry
Word Spotting DTW.
電腦視覺 Computer and Robot Vision I
Light Fields PROPERTIES AND APPLICATIONS. Outline  What are light fields  Acquisition of light fields  from a 3D scene  from a real world scene 
Computer vision: models, learning and inference
University of Pennsylvania 1 GRASP CIS 580 Machine Perception Fall 2004 Jianbo Shi Object recognition.
Computer Vision Detecting the existence, pose and position of known objects within an image Michael Horne, Philip Sterne (Supervisor)
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
Rendering with Concentric Mosaics Heung – Yeung Shum and Li – Wei He Presentation By: Jonathan A. Bockelman.
MSU CSE 240 Fall 2003 Stockman CV: 3D to 2D mathematics Perspective transformation; camera calibration; stereo computation; and more.
Computing motion between images
Flexible Bump Map Capture From Video James A. Paterson and Andrew W. Fitzgibbon University of Oxford Calibration Requirement:
Image-Based Rendering using Hardware Accelerated Dynamic Textures Keith Yerex Dana Cobzas Martin Jagersand.
MSU CSE 803 Fall 2008 Stockman1 CV: 3D sensing and calibration Coordinate system changes; perspective transformation; Stereo and structured light.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Rendering with Concentric Mosaics Heung-Yeung Shum Li-Wei he Microsoft Research.
CSE473/573 – Stereo Correspondence
CSCE 641: Computer Graphics Image-based Rendering Jinxiang Chai.
Stereo Sebastian Thrun, Gary Bradski, Daniel Russakoff Stanford CS223B Computer Vision (with slides by James Rehg and.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
The Story So Far The algorithms presented so far exploit: –Sparse sets of images (some data may not be available) –User help with correspondences (time.
Stereo vision A brief introduction Máté István MSc Informatics.
Path-Based Constraints for Accurate Scene Reconstruction from Aerial Video Mauricio Hess-Flores 1, Mark A. Duchaineau 2, Kenneth I. Joy 3 Abstract - This.
Introduction Tracking the corners Camera model and collision detection Keyframes Path Correction Controlling the entire path of a virtual camera In computer.
Image Based Rendering(IBR) Jiao-ying Shi State Key laboratory of Computer Aided Design and Graphics Zhejiang University, Hangzhou, China
Graphite 2004 Statistical Synthesis of Facial Expressions for the Portrayal of Emotion Lisa Gralewski Bristol University United Kingdom
3D SLAM for Omni-directional Camera
Integral University EC-024 Digital Image Processing.
Reconstructing 3D mesh from video image sequences supervisor : Mgr. Martin Samuelčik by Martin Bujňák specifications Master thesis
03/12/03© 2003 University of Wisconsin Last Time NPR Assignment Projects High-Dynamic Range Capture Image Based Rendering Intro.
Acute angle An angle with a measure less than 90 degrees.
CSCE 643 Computer Vision: Structure from Motion
Image-based Rendering. © 2002 James K. Hahn2 Image-based Rendering Usually based on 2-D imagesUsually based on 2-D images Pre-calculationPre-calculation.
03/24/03© 2003 University of Wisconsin Last Time Image Based Rendering from Sparse Data.
Stereo Many slides adapted from Steve Seitz.
Plenoptic Modeling: An Image-Based Rendering System Leonard McMillan & Gary Bishop SIGGRAPH 1995 presented by Dave Edwards 10/12/2000.
Motion Analysis using Optical flow CIS750 Presentation Student: Wan Wang Prof: Longin Jan Latecki Spring 2003 CIS Dept of Temple.
Vector Graphics Digital Multimedia Chap 이병희
03/09/05© 2005 University of Wisconsin Last Time HDR Image Capture Image Based Rendering –Improved textures –Quicktime VR –View Morphing NPR Papers: Just.
Unit 4 – Flip, Slide & Turn Vocabulary. Unit 4 – Flip, Slide & Turn - Vocabulary Plane: One of the basic undefined terms of geometry. A plane goes on.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
112/5/ :54 Graphics II Image Based Rendering Session 11.
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
High Resolution Surface Reconstruction from Overlapping Multiple-Views
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
776 Computer Vision Jan-Michael Frahm Spring 2012.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Acute angle: An angle with a measure less than 90º.
Eigen Texture Method : Appearance compression based method Surface Light Fields for 3D photography Presented by Youngihn Kho.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Electronic Visualization Laboratory University of Illinois at Chicago “Fast And Reliable Space Leaping For Interactive Volume Rendering” by Ming Wan, Aamir.
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Motion and Optical Flow
COMP 9517 Computer Vision Motion 7/21/2018 COMP 9517 S2, 2012.
Image-Based Rendering
The Brightness Constraint
Mauricio Hess-Flores1, Mark A. Duchaineau2, Kenneth I. Joy3
© 2005 University of Wisconsin
Common Classification Tasks
Coding Approaches for End-to-End 3D TV Systems
The Brightness Constraint
The Brightness Constraint
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Presentation transcript:

Plenoptic Stitching: A Scalable Method for Reconstructing 3D Interactive Walkthroughs Daniel G. Aliaga Ingrid Carlbom Presented by Matthew McCrory

Introduction Interactive walkthroughs require detailed 3D models of their environments Traditional approaches: time consuming, still don’t achieve high detail, subtle lighting effects Plenoptic functions allow capture of high detail, subtle lighting effects, in small amount of time

Plenoptic Functions 7D plenoptic function describes light intensity passing through every viewpoint, in every direction, for all time and wavelengths Currently, all IBR techniques generate lower-dimensional plenoptic functions using a set of images as their input Aliaga’s and Carlbom’s technique reconstructs a 4D plenoptic function

Overview of the process Givens: Viewer moving in open area in a large, complex environment. Camera motion restricted to a plane at eye-height Omnidirectional environment images taken along paths of irregular grid Closed image loops “stitched together” Arbitrary views generated at runtime

Benefits of this Approach Ease of capture – Complex environments captured in a matter of minutes Automated processing – Except camera pose estimations, everything’s automatic Scalability – Method is easily scaled to large environments Support for arbitrarily shaped environments – Because image loops may be irregular, arbitrarily shaped environments can be acquired

Walkthrough Parameterization One possibility: parameterize all potential light rays by their intersection with two perpendicular planes Creates a 4D plenoptic function Vertical field-of-view limited, straight up/straight down ignored

Walkthrough Parameterization To reconstruct a continuous function, the entire open space should be sampled densely, which just isn’t practical! A solution: sample observer plane using irregular grid of omnidirectional image sequences forming image loops in the observer plane

Capture Relatively inexpensive camera system used, built from off-the-shelf components. Camera placed on motorized cart with battery, computer, frame grabber, and fast RAID disk Camera uses a convex paraboloidal mirror w/an orthographic projection

Camera Pose Estimation Calibration scheme developed by Aliaga and Carlbom that uses beacons for calibration In 2 corners of a region, beacons are placed. Before recording, user initializes pose estimation by identifying the projections of the beacons in the first captured image As camera moves, beacon is tracked and, using triangulation, derives camera position and orientation

Reconstruction Given a set of image loops, create novel planar views of environment from arbitrary viewpoints inside a loop Combine pixels from omnidirectional images in the forward-looking view frustum with pixels in the omnidirectional images in the rear-looking frustum

Reconstruction New view created via column-by-column reconstruction of pixels from the omnidirectional images If viewing direction for a column intersects the COP of 2 complimentary omnidirectional images, then that direction maps directly to radial lines in the images.

Reconstruction Corresponding segments of each radial line warped to the column in reconstructed image Geometry of omnidirectional camera must be considered Using similar triangles…

Reconstruction Because pixels from the image behind the viewpoint are generally stretched during the warp, pixels are drawn using fixed sized splats Vertical disocclusions are filled in using longer than expected splats Usually, viewing direction will not intersect exactly with COP of an omnidirectional image Radial is generated then by blending 2 parallel radials

Reconstruction Difficult to reliably identify corresponding features in the radial lines from omnidirectional images on opposite sides of a loop Temporal coherence of the entire image loop is used to identify the required features Start with an arbitrary omni image, track features all the way around the loop

Optimization In ideal conditions, radial lines differ only by a radial displacement and should be easy to recover In practice, feature tracking and camera pose estimation introduce errors into mapping

Optimization: Rotational Correction This optimization accounts for incorrect pairing of radial lines Features should move along same set of circles Error represented by sum of squares of distances between corresponding features Errors reduced by searching for better aligned rotations

Optimization: Column Correlation Ideally, pairs of warped columns should be vertically aligned Feature drifting, lack of sufficient features, and inaccurate distance estimation between the images cause misalignment To correct this, one column is scaled prior to blending the two columns together

Implementation To achieve real-time performance, reconstruction is divided into a preprocessing and a runtime phase A modified JPEG compression and a 3 cache system help to quickly and dynamically load images and data

Results Reconstruction time varies slightly depending on # of features per column Reconstructions of 320x160 and 640x320 pixels take between 5 to 10 frames/sec with a pixel splat of 1x4 pixels Maximum capture time for the four environments recorded was 25 minutes

Results Top image: Reconstructed using described method for viewpoint near middle of a loop Bottom image: Planar re-projection of an image captured from approximately the same viewpoint

Results Example reconstructions for viewpoints near the middle of image loops

Results Environmental statistics:

That’s it! Questions?