1 Computational Vision CSCI 363, Fall 2012 Lecture 18 Stereopsis III.

Slides:



Advertisements
Similar presentations
EXAM 2 !!! Monday, Tuesday, Wednesday, Thursday of NEXT WEEK.
Advertisements

Depth Cues Pictorial Depth Cues: aspects of 2D images that imply depth
COMPUTATIONAL NEUROSCIENCE FINAL PROJECT – DEPTH VISION Omri Perez 2013.
Extra Credit for Participating in Experiments Go to to sign upwww.tatalab.ca We are recruiting for a variety of experiments including basic.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Binocular Disparity points (C) nearer than fixation (P) have crossed disparity points (F) farther than fixation have uncrossed disparity.
The Apparatus. Seeing in Stereo It’s very hard to read words if there are multiple images on your retina.
What is Stereopsis? The process in visual perception that leads to the sensation of depth due to the slightly different perspectives that our two eyes.
Read Pinker article for Thurs.. Seeing in Stereo.
© 2004 by Davi GeigerComputer Vision April 2004 L1.1 Binocular Stereo Left Image Right Image.
Motion Depth Cues – Motion 1. Parallax. Motion Depth Cues – Parallax.
Binocular Disparity points nearer than horopter have crossed disparity
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 8: Stereoscopic Vision 1 Computational Architectures in Biological.
Stereoscopic Depth Disparity between the two retinal images indicates an objects distance from the plane of fixation.
Lecture 4: Edge Based Vision Dr Carole Twining Thursday 18th March 2:00pm – 2:50pm.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
© 2006 by Davi GeigerComputer Vision April 2006 L1.1 Binocular Stereo Left Image Right Image.
Infinity of Interpretations There are an infinite number of interpretations of the 2D pattern of light on the retina.
CSE473/573 – Stereo Correspondence
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Reading Gregory 24 th Pinker 26 th. Seeing Depth What’s the big problem with seeing depth ?
Careers for Psychology and Neuroscience Majors Oct. 19th5-7pm in SU 300 Ballroom B.
1 Computational Vision CSCI 363, Fall 2012 Lecture 26 Review for Exam 2.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
Chapter 5 Human Stereopsis, Fusion, and Stereoscopic Virtual Environments.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Binocular Stereo #1. Topics 1. Principle 2. binocular stereo basic equation 3. epipolar line 4. features and strategies for matching.
CS332 Visual Processing Department of Computer Science Wellesley College Binocular Stereo Vision Region-based stereo matching algorithms Properties of.
Lec 22: Stereo CS4670 / 5670: Computer Vision Kavita Bala.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
1 Computational Vision CSCI 363, Fall 2012 Lecture 5 The Retina.
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities Prof. Charlene Tsai.
1 Perception and VR MONT 104S, Fall 2008 Lecture 4 Lightness, Brightness and Edges.
1 Computational Vision CSCI 363, Fall 2012 Lecture 6 Edge Detection.
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry Topics: Basics of projective geometry Points and hyperplanes in projective space Homography.
Computational Vision CSCI 363, Fall 2012 Lecture 22 Motion III
Outline Of Today’s Discussion 1.Monocular & Binocular Depth Cues: Understanding Retinal Disparity.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
Outline Of Today’s Discussion 1.Some Disparities are Not Retinal: Pulfrich Effect 2.Random-Dot Stereograms 3.Binocular Rivalry 4.Motion Parallax.
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities May 2, 2005 Prof. Charlene Tsai.
1 Computational Vision CSCI 363, Fall 2012 Lecture 29 Structure from motion, Heading.
Computational Vision CSCI 363, Fall 2012 Lecture 17 Stereopsis II
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
ORTH 140 NORMAL BINOCULAR SINGLE VISION AND MOTOR FUSION
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
Journal of Vision. 2011;11(8):11. doi: / Figure Legend:
Digital Image Processing Lecture 16: Segmentation: Detection of Discontinuities Prof. Charlene Tsai.
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
STEREOPSIS The Stereopsis Problem: Fusion and Reconstruction
Space Perception and Binocular Vision
Range Imaging Through Triangulation
Binocular Stereo Vision
Binocular Stereo Vision
Binocular Stereo Vision
Binocular Disparity and the Perception of Depth
Detecting image intensity changes
Course 6 Stereo.
Binocular Stereo Vision
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Receptive Fields of Disparity-Tuned Simple Cells in Macaque V1
Binocular Stereo Vision
Detection of image intensity changes
Presentation transcript:

1 Computational Vision CSCI 363, Fall 2012 Lecture 18 Stereopsis III

2 Summary of Marr-Poggio Horizontal lines: constant Right eye position. Vertical lines: Constant left eye position. Diagonal lines: Constant disparity.

3 Limitations of the Marr-Poggio Algorithm Problems with Marr-Poggio algorithm: Does not specify what features are being matched. Does not make use of zero crossings or multiple spatial scales. Does not make use of vergence eye movements. Importance of Vergence Eye Movements: Humans can only fuse images with disparities of about +/- 10 min of arc. This range is called Panum's fusional area. If we need to fuse the images of an object that is nearer or farther than this, we make vergence eye movements to change the plane of fixation. Humans make many vergence eye movements as we scan the environment.

4 Marr-Poggio-Grimson Basic idea: Use zero crossings after filtering with Large leads to widely spread zero crossings that are easier to match but may give less accurate estimates of disparity. Small leads to more finely tuned disparities. Start with large filters, and find matches for the zero crossings. We can use the estimates of disparity with the large filters to guide vergence eye movements to align features for comparing finer disparities with the smaller filters.

5 Zero crossings for different filter sizes

6 Range of disparities to search Suppose we filter an image with a operator for which the width of the inner portion is w. Given a zero crossing at some position in the filtered image, the probability of another zero crossing within +/- w/2 of the first one is 5%. Therefore, if there is a zero crossing at a given position in the left image and a zero crossing in the right image within the range of +/- w/2 of the one in the left image, the probability that these two are a correct match is 95%. +w/2 -w/2 L R d

7 Determining correctness If the correct disparity is in the range of +/- w/2, then we will find matches for nearly all the zero crossings in the left image (95%). If the correct disparity is not in the range, then the probability of a match is 40%. Therefore, if we are only finding matches for 40% of the zero crossings, we are not looking at the correct range of disparities. We can correct this by making vergence eye movements.

8 Increasing the range of disparities examined The actual algorithm uses a range of +/- w. This leads to more false matches. For a range of +/- w, the likelihood of a false match is at most 50%. So 50% of the time our matches will be correct. If we have a false match, we usually have one crossed disparity and one uncrossed disparity matches. We can use neighboring points to disambiguate the false matches.

9 Disambiguating false matches To determine the correct matches, we match both from the left image to the right and from the right image to the left. L1L1 L2L2 R1R1 R2R2 L 2 can only match R 2 R 1 can only match L 1 These determine the matches uniquely. It can be shown that if more than 70% of the zero crossings have a match within +/- w, then the disparity range is correct. If not, use a vergence eye movement and try again.

10 The full algorithm A.To determine initial correspondence: 1. Find zero crossings using a operator with a central width, w. 2. For each horizontal slice: 2.1 Find the nearest neighbors in the right image for each zero- crossing in the left image. 2.2 Find the nearest neighbors in the left image for each zero- crossing in the right image. 2.3 For each pair of zero-crossings that are closest neighbors of one another, find the distance,  initial, of the right fragment from the left. If  initial is less than the matching tolerance, m, consider the zero crossings matched with disparity  initial.

11 Finishing the algorithm B. To determine final correspondence: 1.Find the zero-crossings using a operator of width w/2. 2.For each horizontal slice: 2.1 For each zero crossing in the left image: Find the nearest zero crossing in the left image that matched when the operator was width w Offset the zero crossing by a distance of  initial, the disparity of the nearest matching zero crossing found at the lower resolution with operator width w. 2.2 Find the nearest neighbors in the right image for the zero crossing in the left image 2.3 Find the nearest neighbors in the left image for the zero crossing in the right image. 2.4 For each pair of zero crossings that are closest neighbors to one another, find the distance,  new, that the right is separated from the left. If  new is less than the matching tolerance, m/2, then consider the zero crossings matched with a total disparity of  final =  new +  initial.

12 Do Humans Use the Same Constraints as Marr-Poggio? 1.Similarity: We probably use this one. Humans cannot fuse a white dot with a black dot. 2.Epipolar: The brain probably uses some version of this. If one image is shifted upward (or downward), people cannot fuse the two images. 3.Uniqueness: We probably don't rely on this. There are examples of images where can fuse two features in one image with a single feature in the other.

13 Violations of Uniqueness constraint Panum's limiting case: Matching one line with two Braddick's demonstration: Matching one point with two repeatedly Stereo algorithms can deal with Braddick's demonstration with slight modifications.

14 The Continuity Constraint The brain probably uses some form of continuity constraint. Evidence: There is a limit to how quickly disparity can change from one location to the next, and still produce stereo fusion. For a plane that is steeply slanted in depth, people lose the ability to see the slant and just see the step edge.

15 Does the Brain use Zero Crossings? Many machine vision algorithms extract edges first (e.g. with zero crossings) and then compute the disparities of matched edges. They use edges for matching because they correspond to important physical features (e.g. object boundaries). We also know that people can localize the positions of edges very accurately. This accurate localization is required for stereo vision. However, it is not clear what primitives the human brain matches when computing stereo disparity. The information is combined in V1, so it is after the center- surround operators (like the laplacian operators). Zero crossing information would be available.

16 Zero crossings are not enough Stereo Pairs Luminance Profile Convolved images Left Right

17 Perception vs. Computation Perceived Depth Computed depth of zero crossings Positions of peaks and troughs Computed depth using peaks and troughs

18 Some V1 cells are tuned to disparity Tuned Excitatory Tuned Inhibitory Some cells are narrowly tuned for disparity. Most prefer a disparity near zero.

19 Near and Far cells Some cells are broadly tuned for disparity, preferring either near objects or far objects.