John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Stereo Vision Iolanthe in the Bay.

Slides:



Advertisements
Similar presentations
Stereo Vision Reading: Chapter 11
Advertisements

CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
Public Library, Stereoscopic Looking Room, Chicago, by Phillips, 1923.
Stereo Many slides adapted from Steve Seitz. Binocular stereo Given a calibrated binocular stereo pair, fuse it to produce a depth image Where does the.
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Stereo.
3D Computer Vision and Video Computing 3D Vision Topic 3 of Part II Stereo Vision CSc I6716 Spring 2011 Zhigang Zhu, City College of New York
Last Time Pinhole camera model, projection
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2005 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
1 Introduction to 3D Imaging: Perceiving 3D from 2D Images How can we derive 3D information from one or more 2D images? There have been 2 approaches: 1.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
Stereopsis Mark Twain at Pool Table", no date, UCR Museum of Photography.
The plan for today Camera matrix
CS 223b 1 More on stereo and correspondence. CS 223b 2 =?f g Mostpopular For each window, match to closest window on epipolar line in other image. (slides.
3D from multiple views : Rendering and Image Processing Alexei Efros …with a lot of slides stolen from Steve Seitz and Jianbo Shi.
3D Computer Vision and Video Computing 3D Vision Topic 4 of Part II Stereo Vision CSc I6716 Spring 2008 Zhigang Zhu, City College of New York
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
3D Computer Vision and Video Computing 3D Vision Lecture 15 Stereo Vision (II) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
CSE473/573 – Stereo Correspondence
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
COMP322/S2000/L271 Stereo Imaging Ref.V.S.Nalwa, A Guided Tour of Computer Vision, Addison Wesley, (ISBN ) Slides are adapted from CS641.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Stereo matching “Stereo matching” is the correspondence problem –For a point in Image #1, where is the corresponding point in Image #2? C1C1 C2C2 ? ? C1C1.
Stereo vision A brief introduction Máté István MSc Informatics.
1 Perceiving 3D from 2D Images How can we derive 3D information from one or more 2D images? There have been 2 approaches: 1. intrinsic images: a 2D representation.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
John Morris Iolanthe returns to the Waitemata Harbour Stereo Vision The Correspondence Problem.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Structure from images. Calibration Review: Pinhole Camera.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Stereo Vision Iolanthe in the Bay.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
Stereo Vision Reading: Chapter 11 Stereo matching computes depth from two or more images Subproblems: –Calibrating camera positions. –Finding all corresponding.
3D Sensing and Reconstruction Readings: Ch 12: , Ch 13: , Perspective Geometry Camera Model Stereo Triangulation 3D Reconstruction by.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
Stereo Many slides adapted from Steve Seitz.
#? rahul swaminathan (T-Labs) & professor patrick baudisch hci2 hasso-plattner institute determining depth.
Acquiring 3D models of objects via a robotic stereo head David Virasinghe Department of Computer Science University of Adelaide Supervisors: Mike Brooks.
Computer Vision, Robert Pless
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Computer Vision Stereo Vision. Bahadir K. Gunturk2 Pinhole Camera.
Stereo Vision John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Vision Research in.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
Lecture 16: Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
stereo Outline : Remind class of 3d geometry Introduction
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
Stereo Vision John Morris Vision Research in CITR
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Computer vision: models, learning and inference M Ahad Multiple Cameras
3D Reconstruction Using Image Sequence
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Computer vision: geometric models Md. Atiqur Rahman Ahad Based on: Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Stereo CS4670 / 5670: Computer Vision Noah Snavely Single image stereogram, by Niklas EenNiklas Een.
제 5 장 스테레오.
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Motion and Optical Flow
Common Classification Tasks
Thanks to Richard Szeliski and George Bebis for the use of some slides
Computer Vision Stereo Vision.
Course 6 Stereo.
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Presentation transcript:

John Morris These slides were adapted from a set of lectures written by Mircea Nicolescu, University of Nevada at Reno Stereo Vision Iolanthe in the Bay of Islands

2 Stereo Vision Goal −Recovery of 3D scene structure −Using two or more images, −Each acquired from a different viewpoint in space −Using multiple cameras or one moving camera −Term binocular vision is used when two cameras are employed −More than 2 cameras can be used  Acquisition of complete 3D models Stereophotogrammetry Using stereo vision systems to measure properties (dimensions here) of a scene

3 Stereo Vision - Terminology −Fixation point −Point of intersection of the optical axes of the two cameras −Baseline −Distance between the camera optical centres −Epipolar plane −Plane passing through the optical centres and a point in the scene −Epipolar line −Intersection of the epipolar plane with the image plane. −Conjugate pair or Corresponding points −A point in the scene visible to both cameras (binocularly visible) will be projected to a point in each image −Disparity −Distance between corresponding points when the two images are superimposed −Disparity map −Disparities of all points form the disparity map −Usual output from a stereo matching algorithm −Often displayed as an image

4 Stereo Vision Camera configuration Parallel optical axes Parallel image planes Note: Virtual Image planes (in front of optical centre)

5 Stereo Vision – Verging axes Camera configuration Verging optical axes

6 Triangulation Principle underlying stereo vision Any visible point in the scene must lie on the line that passes through −the optical centre (centre of projection) and −the projection of the point on the image plane We can backproject this line into the scene With two cameras, we have two such lines Intersection of these two lines is the (3D) location of the point

7 Stereo Vision Two problems −Correspondence problem −Reconstruction problem Correspondence problem −Finding conjugate pairs of corresponding or matched points in each image −These points are projections of the same scene point −Triangulation depends on these conjugate pairs

8 Stereo Vision Correspondence problem −Ambiguous correspondence between points in the two images may lead to several different consistent interpretations of the scene −Problem is fundamentally ill-posed If you can’t solve the correspondence problem, then all of these points could be scene points! Each image has 3 scene points, representing some features in the scene

9 Reconstruction −Having found the corresponding points, we can compute the disparity map −Disparity maps are commonly expressed in pixels ie number of pixels between corresponding points in two images −Disparity map can be converted to a 3D map of the scene if the geometry of the imaging system is known −Critical parameters: Baseline, camera focal length, pixel size

10 Reconstruction Determining depth −To recover the position of P from its projections, p l and p r : −In general, the two cameras are related by a rotation, R, and a translation, T : −Parallel camera optical axes  Z r = Z l = Z and X r = X l – T so we have: where d = x l – x r is the disparity - the difference in position between the corresponding points in the two images, commonly measured in pixels

11 Reconstruction Recovering depth where T is the baseline If d’ is measured in pixels, then d = x l – x r = d’p where p is the width of a pixel in the image plane, and we have Z = Tf Note the reciprocal relationship between disparity and depth! This is particularly relevant when considering the accuracy of stereo photogrammetry d’p

12 Stereo Vision Configuration parameters −Intrinsic parameters −Characterize the transformation from image plane coordinates to pixel coordinates in each camera −Parameters intrinsic to each camera −Extrinsic parameters ( R, T ) −Describe the relative position and orientation of the two cameras −Can be determined from the extrinsic parameters of each camera:

13 Correspondence Problem Why is the correspondence problem difficult? −Some points in each image will have no corresponding points in the other image −They are not binocularly visible or −They are only monocularly visible −Cameras have different fields of view − Occlusions may be present −A stereo system must be able to determine parts that should not be matched These two are equivalent!

14 The Correspondence Problem Methods for establishing correspondences −Two issues −How to select candidate matches? −How to determine the goodness of a match? −Two main classes of correspondence (matching) algorithm: −Correlation-based −Attempt to establish a correspondence by matching image intensities – usually over a window of pixels in each image  Dense disparity maps −Distance is found for all BV image points −Except occluded (MV) points −Feature-based −Attempt to establish a correspondence by matching a sparse sets of image features – usually edges −Disparity map is sparse −Number of points is related to the number of image features identified

15 Correlation-Based Methods Match image sub-windows in the two images using image correlation −oldest technique for finding correspondence between image pixels Scene points must have the same intensity in each image − Assumes a)All objects are perfect Lambertian scatterers ie the reflected intensity is not dependent on angle or objects scatter light uniformly in all directions Informally - matte surfaces only b)Fronto-planar surfaces −(Visible) surfaces of all objects are perpendicular to camera optical axes

16 Correlation-Based Methods

17 Correlation-Based Methods Usually, we normalize c(d) by dividing it by the standard deviation of both I l and I r (normalized cross-correlation, c(d)  [0,1] ) where and are the average pixel values in the left and right windows. An alternative similarity measure is the sum of squared differences (SSD): In fact, experiment shows that the simpler sum of absolute differences (SAD) is just as good c(d) =   | I l ( i+k, j+l ) – I r ( i+k-d, j+l ) |

18 Correlation-Based Methods Improvements −Instead of using the image intensity values, the accuracy of correlation is improved by using thresholded signed gradient magnitudes at each pixel. −Compute the gradient magnitude at each pixel in the two images without smoothing −Map the gradient magnitude values into three values: -1, 0, 1 (by thresholding the gradient magnitude) −More sensitive correlations are produced this way + several dozen more see Scharstein & Szeliski, 2001 for a review

19 Correlation-Based Methods Comments −Correlation-based methods depend on the image window in one image having a distinctive structure that occurs infrequently in the search region of the other image ie in one image, we can find unique features in each window that match only one window in the other −How to choose the size of the window, W ? −too small −may not capture enough image structure and −may be too noise sensitive  many false matches −too large −makes matching less sensitive to noise (desired) but −decreases precision (blurs disparity map) −An adaptive searching window has been proposed

20 Correlation-Based Methods

21 Correlation-Based Methods

22 Correlation-Based Methods Comments −How to choose the size and location of the search region, R(p l )? −if the distance of the fixating point from the cameras is much larger than the baseline, the location of R(p l ) can be chosen to be the same as the location of p l −the size (extent) of R(p l ) can be estimated from the maximum range of distances we expect to find in the scene −we will see that the search region can always be reduced to a line

23 Feature-Based Methods Main idea −Look for a feature in an image that matches a feature in the other. −Typical features used are: −edge points −line segments −corners (junctions)

24 Feature-Based Methods A set of features is used for matching −a line feature descriptor, for example, could contain: −length, l −orientation,  −coordinates of the midpoint, m −average intensity along the line, i Similarity measures are based on matching feature descriptors: where w 0,..., w 3 are weights (determining the weights that yield the best matches is a nontrivial task).

25 Feature-Based Methods

26 Correlation vs. feature-based approaches Correlation methods −Easier to implement −Provide a dense disparity map (useful for reconstructing surfaces) −Need textured images to work well (many false matches otherwise) −Don’t work well when viewpoints are very different, due to −change in illumination direction −violates Lambertian scattering assumption −foreshortening −perspective problem – surfaces are not fronto-planar Feature-based methods −Suitable when good features can be extracted from the scene −Faster than correlation-based methods −Provide sparse disparity maps −OK for applications like visual navigation −Relatively insensitive to illumination changes

27 Other correspondence algorithms Dynamic programming (Gimel’Farb) −Finds a ‘path’ through an image which provides the best (least-cost) match −Can allow for occlusions (Birchfield and Tomasi) −Generally provide better results than area-based correlation −Faster than correlation Graph Cut (Zabih et al) −Seems to provide best results −Very slow, not suitable for real-time applications Concurrent Stereo Matching −Examine all possible matches in parallel (Delmas, Gimel’Farb, Morris, work in progress ) −Uses a model of image noise instead of arbitrary weights in cost functions −Suitable for real-time parallel hardware implementation Some of these will be considered in detail later