Common Classification Tasks

Slides:



Advertisements
Similar presentations
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Advertisements

CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
Computer Vision Lecture 16: Texture
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2005 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
1 Introduction to 3D Imaging: Perceiving 3D from 2D Images How can we derive 3D information from one or more 2D images? There have been 2 approaches: 1.
Introduction to Computer Vision 3D Vision Topic 9 Stereo Vision (I) CMPSCI 591A/691A CMPSCI 570/670.
3D from multiple views : Rendering and Image Processing Alexei Efros …with a lot of slides stolen from Steve Seitz and Jianbo Shi.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
© 2004 by Davi GeigerComputer Vision March 2004 L1.1 Binocular Stereo Left Image Right Image.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Lec 21: Fundamental Matrix
CSE473/573 – Stereo Correspondence
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Stereo vision A brief introduction Máté István MSc Informatics.
Project 4 Results Representation – SIFT and HoG are popular and successful. Data – Hugely varying results from hard mining. Learning – Non-linear classifier.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Reflectance Map: Photometric Stereo and Shape from Shading
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
CAP4730: Computational Structures in Computer Graphics 3D Concepts.
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
1 Preview At least two views are required to access the depth of a scene point and in turn to reconstruct scene structure Multiple views can be obtained.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Stereo Course web page: vision.cis.udel.edu/~cv April 11, 2003  Lecture 21.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
1 Formation et Analyse d’Images Session 7 Daniela Hall 25 November 2004.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Lecture 03 15/11/2011 Shai Avidan הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
stereo Outline : Remind class of 3d geometry Introduction
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Computer vision: models, learning and inference M Ahad Multiple Cameras
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Computer vision: geometric models Md. Atiqur Rahman Ahad Based on: Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Another Example: Circle Detection
MAN-522 Computer Vision Spring
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
Processing visual information for Computer Vision
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth?
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
Three-Dimensional Concepts. Three Dimensional Graphics  It is the field of computer graphics that deals with generating and displaying three dimensional.
Computer Vision Lecture 4: Color
Fitting Curve Models to Edges
Range Imaging Through Triangulation
Computer Vision Lecture 3: Digital Images
Computer Vision Lecture 16: Texture II
Course 6 Stereo.
Introduction to Artificial Intelligence Lecture 24: Computer Vision IV
Fourier Transform of Boundaries
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Shape from Shading and Texture
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Common Classification Tasks Recognition of individual objects/faces Analyze object-specific features (e.g., key points) Train with images from different viewing angles Recognition of object classes Analyze features that are consistent within class and differ between them as much as possible. Train with many exemplars from each class. Recognition of scene types Find and analyze common features, objects, or layouts within scene classes. Use large variety of scene photos. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Example: AlexNet April 26, 2018 Computer Vision Lecture 21: Depth

Another Example: Inception April 26, 2018 Computer Vision Lecture 21: Depth

Another Example: Inception April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Estimating the distance of a point from the observer is crucial for scene and object recognition. There are many monocular cues such as shading, texture gradient, and perspective. However, just like biological systems, computer vision systems can benefit a lot from stereo vision, i.e. using two visual inputs from slightly different angles. Computer vision systems also have the option to use completely different approaches such as radar, laser range finding, and structured lighting. Let us first talk about stereo vision. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision In the simplest case, the two cameras used for stereo vision are identical and separated only along the x-axis by a baseline distance b. Then the image planes are coplanar. Depth information is obtained through the fact that the same feature point in the scene appears at slightly different positions in the two image planes. This displacement between the two images is called the disparity. The plane spanned by the two camera centers and the feature point is called the epipolar plane. Its intersection with the image plane is the epipolar line. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision Geometry of binocular stereo vision April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision Ideally, every feature in one image will lie in the same vertical position in the other image. However, due to distortion in the camera images and imperfect geometry, there usually is also some vertical disparity. While many algorithms for binocular stereo do not account for such vertical disparity, it increases the robustness of our system if we allow at least a few pixels of deviation from the epipolar line. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision Before we discuss the computation of depth from stereo imaging, let us look at two definitions: A conjugate pair is two points in different images that are the projections of the same point in the scene. Disparity is the distance between points of a conjugate pair when the two images are superimposed. Let us say that the scene point P is observed at points pl and pr in the left and right image planes, respectively. For convenience, let us assume that the origin of the coordinate system coincides with the left lens center. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision Comparing the similar triangles PMCl and plLCl, we get: Similarly, for PNCr and prRCr, we get: Combining gives us: April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Vision Due to the limited resolution of images, increasing the baseline distance b gives us a more precise estimate of depth z. However, the greater b, the more different are the two viewing angles, and the more difficult it can become to determine the correspondence between the two images. This brings us to the main problem in stereo vision: How can we find the conjugate pairs in our stereo images? This problem is called stereo matching. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching In stereo matching, we have to solve a problem that is still under investigation by many researchers, called the correspondence problem. It can be phrased like this: For each point in the left image, find the corresponding point in the right image. The idea underlying all stereo matching algorithms is that these two points should be similar to each other. So we need a measure for similarity. Moreover, we need to find matchable features. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Stereo Matching April 26, 2018 Computer Vision Lecture 21: Depth

Generating Interesting Points Using interpolation to determine the depth of points within large homogeneous areas may cause large errors. To overcome this problem, we can generate additional interesting points that can be matched between the two images. The idea is to use structured light, i.e., project a pattern of light onto the visual scene. This creates additional variance in the brightness of pixels and increases the number of interesting points. April 26, 2018 Computer Vision Lecture 21: Depth

Generating Interesting Points April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Shape from Shading Besides binocular disparity, there are many different ways of depth estimation based on monocular information. For example, if we know the reflective properties of the surfaces in our scene and the position of the light source, we can use shape from shading techniques: Basically, since the amount of light reflected by a surface depends on its angle towards the light source, we can estimate the orientation of surfaces based on their intensity. More sophisticated methods also use the contours of shadows cast by objects to estimate the shape and orientation of those objects. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Photometric Stereo To improve the coarse estimates of orientation derived from shape from shading methods, we can use photometric stereo. This technique uses three light sources that are located at different, known positions. Three images are taken, one for each light source, with the other two light sources being turned off. This way we determine three different intensities for each surface in the scene. These three values put narrow constraints on the possible orientation of a surface and allow a rather precise estimation. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Shape from Texture As we discussed before, the texture gradient gives us valuable monocular depth information. At any point in the image showing texture, the texture gradient is a two-dimensional vector pointing towards the steepest increase in the size of texture elements. The texture gradient across a surface allows a good estimate of the spatial orientation of that surface. Of course, it is important for this technique that the image has high resolution and a precise method of texture size (granularity) measurement is used. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Shape from Motion The shape from motion technique s similar to binocular stereo, but it uses only one camera. This camera is moved while it takes images from the visual scene. This way two images with a specific baseline distance can be obtained, and depth can be computed just like for binocular stereo. We can even use more than two images in this computation to get more robust measurements of depth. The disadvantages of shape from motion techniques are the technical overhead for moving the camera and the reduced temporal resolution. April 26, 2018 Computer Vision Lecture 21: Depth

Computer Vision Lecture 21: Depth Range Imaging If we can determine the depth of every pixel in an image, we can make this information available in the form of a range image. A range image has exactly the same size and number of pixels as the original image. However, each pixel does not specify color or intensity, but the depth of that pixel in the original image, encoded as grayscale. Usually, the brighter a pixel in a range image is, the closer the corresponding pixel is to the observer. By providing both the original and the range image, we basically define a 3D image. April 26, 2018 Computer Vision Lecture 21: Depth

Sample Range Image of a Mug April 26, 2018 Computer Vision Lecture 21: Depth

Range Imaging Through Triangulation How can we obtain precise depth information for every pixel in the image of a scene? One precise but slow method uses a laser that can rotate around its vertical axis and can also assume different vertical positions. This laser systematically and sequentially illuminates points in the image. A scene camera determines the position of every single point in its picture. The trick is that this camera looks at the scene from a different direction than does the laser pointer. Therefore, the depth of every point can be easily and precisely determined through triangulation: April 26, 2018 Computer Vision Lecture 21: Depth

Range Imaging Through Triangulation April 26, 2018 Computer Vision Lecture 21: Depth