December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.

Slides:



Advertisements
Similar presentations
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Advertisements

Image Analysis Phases Image pre-processing –Noise suppression, linear and non-linear filters, deconvolution, etc. Image segmentation –Detection of objects.
CS 376b Introduction to Computer Vision 04 / 21 / 2008 Instructor: Michael Eckmann.
September 10, 2013Computer Vision Lecture 3: Binary Image Processing 1Thresholding Here, the right image is created from the left image by thresholding,
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
1 Lecture 11 Scene Modeling. 2 Multiple Orthographic Views The easiest way is to project the scene with parallel orthographic projections. Fast rendering.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
© 2004 by Davi GeigerComputer Vision March 2004 L1.1 Binocular Stereo Left Image Right Image.
© 2006 by Davi GeigerComputer Vision April 2006 L1.1 Binocular Stereo Left Image Right Image.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Lec 21: Fundamental Matrix
CSE473/573 – Stereo Correspondence
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Visualization- Determining Depth From Stereo Saurav Basu BITS Pilani 2002.
COMP322/S2000/L271 Stereo Imaging Ref.V.S.Nalwa, A Guided Tour of Computer Vision, Addison Wesley, (ISBN ) Slides are adapted from CS641.
Stereo vision A brief introduction Máté István MSc Informatics.
Project 4 Results Representation – SIFT and HoG are popular and successful. Data – Hugely varying results from hard mining. Learning – Non-linear classifier.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Image processing Lecture 4.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
MIND: The Cognitive Side of Mind and Brain  “… the mind is not the brain, but what the brain does…” (Pinker, 1997)
Lecture 12 Stereo Reconstruction II Lecture 12 Stereo Reconstruction II Mata kuliah: T Computer Vision Tahun: 2010.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
CIS 601 Fall 2003 Introduction to Computer Vision Longin Jan Latecki Based on the lectures of Rolf Lakaemper and David Young.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
Shape from Stereo  Disparity between two images  Photogrammetry  Finding Corresponding Points Correlation based methods Feature based methods.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
Stereo Course web page: vision.cis.udel.edu/~cv April 11, 2003  Lecture 21.
November 13, 2014Computer Vision Lecture 17: Object Recognition I 1 Today we will move on to… Object Recognition.
Course 9 Texture. Definition: Texture is repeating patterns of local variations in image intensity, which is too fine to be distinguished. Texture evokes.
1 Formation et Analyse d’Images Session 7 Daniela Hall 25 November 2004.
Computer Vision Michael Isard and Dimitris Metaxas.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
1 Artificial Intelligence: Vision Stages of analysis Low level vision Surfaces and distance Object Matching.
CSE 185 Introduction to Computer Vision Stereo. Taken at the same time or sequential in time stereo vision structure from motion optical flow Multiple.
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
(c) 2000, 2001 SNU CSE Biointelligence Lab Finding Region Another method for processing image  to find “regions” Finding regions  Finding outlines.
Visual Perception Principles Visual perception principles are ‘rules’ that we apply to visual information to assist our organisation and interpretation.
Computer Science Readings: Reinforcement Learning Presentation by: Arif OZGELEN.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
Computer vision: models, learning and inference M Ahad Multiple Cameras
ECE 638: Principles of Digital Color Imaging Systems Lecture 4: Chromaticity Diagram.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Correspondence and Stereopsis Original notes by W. Correa. Figures from [Forsyth & Ponce] and [Trucco & Verri]
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
1 Computational Vision CSCI 363, Fall 2012 Lecture 18 Stereopsis III.
Copyright © Cengage Learning. All rights reserved. 1 Functions and Limits.
Correspondence and Stereopsis. Introduction Disparity – Informally: difference between two pictures – Allows us to gain a strong sense of depth Stereopsis.
A Plane-Based Approach to Mondrian Stereo Matching
55:148 Digital Image Processing Chapter 11 3D Vision, Geometry
Processing visual information for Computer Vision
Think-Pair-Share What visual or physiological cues help us to perceive 3D shape and depth?
Computational Vision CSCI 363, Fall 2016 Lecture 15 Stereopsis
Common Classification Tasks
Computer Vision Lecture 5: Binary Image Processing
Overview Pin-hole model From 3D to 2D Camera projection
Perception We have previously examined the sensory processes by which stimuli are encoded. Now we will examine the ultimate purpose of sensory information.
Object Recognition Today we will move on to… April 12, 2018
Course 6 Stereo.
Fourier Transform of Boundaries
Chapter 11: Stereopsis Stereopsis: Fusing the pictures taken by two cameras and exploiting the difference (or disparity) between them to obtain the depth.
Presentation transcript:

December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding

December 2, 2014Computer Vision Lecture 21: Image Understanding 2 Image Understanding Machine vision consists of lower and upper processing levels. Image understanding is the highest processing level in this classification. The main computer vision goal is to achieve machine behavior similar to that of biological systems by applying technically available procedures. Let us look at both technical and biological systems and how they achieve image understanding.

December 2, 2014Computer Vision Lecture 21: Image Understanding 3 Control Strategies Let us take a look at different Image understanding control strategies in computer vision, which are actually similar to biological principles. Parallel and serial processing control: Parallel processing makes several computations simultaneously. Serial processing operations are sequential. Almost all low-level image processing can be done in parallel. High-level processing using higher levels of abstraction is usually serial in essence.

December 2, 2014Computer Vision Lecture 21: Image Understanding 4 Control Strategies Hierarchical control: Control by the image data (bottom-up control): Processing proceeds from the raster image to segmented image, to region (object) description, and to recognition.

December 2, 2014Computer Vision Lecture 21: Image Understanding 5 Control Strategies Hierarchical control (continued): Model-based control (top-down control): A set of assumptions and expected properties is constructed from applicable knowledge. The satisfaction of those properties is tested in image representations at different processing levels in top- down direction, down to the original image data. The image understanding is an internal model verification, and the model is either accepted or rejected.

December 2, 2014Computer Vision Lecture 21: Image Understanding 6 Control Strategies Hierarchical control (continued): Combined control uses both data driven and model driven control strategies. Non-hierarchical control does not distinguish between upper and lower processing levels. It can be seen as a cooperation of competing experts at the same level (often) using the blackboard principle. The blackboard is a shared data structure that can be accessed by multiple experts.

December 2, 2014Computer Vision Lecture 21: Image Understanding 7 Control Strategies Interestingly, human vision also relies on both data driven (bottom-up) and model driven (or task driven, top-down) control strategies for visual attention. Such control strategies can be studied by, for example, conducting visual search experiments. For a demonstration, try to determine on the following slide whether it contains the letter “O”.

December 2, 2014Computer Vision Lecture 21: Image Understanding 8 Visual Search

December 2, 2014Computer Vision Lecture 21: Image Understanding 9 Visual Search That was not very difficult, right? In fact, even without the task to find an “O”, your attention would have been drawn to the “O” immediately. Because it is the only “odd” item in the display, the “O” is most conspicuous (salient). Bottom-up control of attention will therefore make you look at the “O” instantly; no top-down controlled search process is necessary. On the next slide, try to determine whether the display contains an orange square.

December 2, 2014Computer Vision Lecture 21: Image Understanding 10 Visual Search

December 2, 2014Computer Vision Lecture 21: Image Understanding 11 Visual Search This time, the target of your search most likely did not “pop out” at you, but you needed to search through the display with the orange square on your mind. In this case, top-down (model- or task-driven) control of your visual attention was necessary for task completion. On the following slide, try to find out whether it contains a black “O”.

December 2, 2014Computer Vision Lecture 21: Image Understanding 12 Visual Search

December 2, 2014Computer Vision Lecture 21: Image Understanding 13 Visual Search This task was easier again. However, you may have been distracted, at least for the fraction of a second, by the red X in the display. You knew that both its color and its shape were irrelevant to your task. Thus, in this case, your bottom-up control interfered with your top-down control.

December 2, 2014Computer Vision Lecture 21: Image Understanding 14 Visual Search In us humans, the interaction of bottom-up and top-down control is tuned in an adequate way for everyday life. Top-down control is used whenever we need to do a particular visual task. Bottom-up control allows us to recognize conspicuous – possibly dangerous – information quickly. Therefore, it is necessary that bottom-up control can sometimes override top-down control. In the previous example, this overriding rather impeded our search task. However, in everyday life (traffic, for example), it can sometimes prevent dangerous situations.

December 2, 2014Computer Vision Lecture 21: Image Understanding 15 Control Strategies What do you see in this picture?

December 2, 2014Computer Vision Lecture 21: Image Understanding 16 Control Strategies At first you may just have seen an abstract black-and-white pattern. After some time, you may have recognized a dog (Dalmatian). From now on, even in a few weeks from now, if you look at the picture again, you will immediately see the dog. This shows how important model- and hypothesis driven (top- down) processes are for our vision. In our visual system, bottom-up and top-down processes continuously interact to derive the best possible interpretation of the current scene in our visual field. Ideally, a computer vision system should do the same to achieve powerful image understanding capabilities.

December 2, 2014Computer Vision Lecture 21: Image Understanding 17 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 18 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 19 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 20 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 21 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 22 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 23 Control Strategies

December 2, 2014Computer Vision Lecture 21: Image Understanding 24 Depth

December 2, 2014Computer Vision Lecture 21: Image Understanding 25Depth Estimating the distance of a point from the observer is crucial for scene and object recognition. There are many monocular cues such as shading, texture gradient, and perspective. However, just like biological systems, computer vision systems can benefit a lot from stereo vision, i.e. using two visual inputs from slightly different angles. Computer vision systems also have the option to use completely different approaches such as radar, laser range finding, and structured lighting. Let us first talk about stereo vision.

December 2, 2014Computer Vision Lecture 21: Image Understanding 26 Stereo Vision In the simplest case, the two cameras used for stereo vision are identical and separated only along the x- axis by a baseline distance b. Then the image planes are coplanar. Depth information is obtained through the fact that the same feature point in the scene appears at slightly different positions in the two image planes. This displacement between the two images is called the disparity. The plane spanned by the two camera centers and the feature point is called the epipolar plane. Its intersection with the image plane is the epipolar line.

December 2, 2014Computer Vision Lecture 21: Image Understanding 27 Stereo Vision Geometry of binocular stereo vision

December 2, 2014Computer Vision Lecture 21: Image Understanding 28 Stereo Vision Ideally, every feature in one image will lie in the same vertical position in the other image. However, due to distortion in the camera images and imperfect geometry, there usually is also some vertical disparity. While many algorithms for binocular stereo do not account for such vertical disparity, it increases the robustness of our system if we allow at least a few pixels of deviation from the epipolar line.

December 2, 2014Computer Vision Lecture 21: Image Understanding 29 Stereo Vision Before we discuss the computation of depth from stereo imaging, let us look at two definitions: A conjugate pair is two points in different images that are the projections of the same point in the scene. Disparity is the distance between points of a conjugate pair when the two images are superimposed. Let us say that the scene point P is observed at points p l and p r in the left and right image planes, respectively. For convenience, let us assume that the origin of the coordinate system coincides with the left lens center.