Introduction to Robot Vision

Slides:



Advertisements
Similar presentations
Lecture 11: Two-view geometry
Advertisements

Last 4 lectures Camera Structure HDR Image Filtering Image Transform.
CS 691 Computational Photography Instructor: Gianfranco Doretto 3D to 2D Projections.
Study the mathematical relations between corresponding image points.
Computer vision: models, learning and inference
Camera Calibration. Issues: what are intrinsic parameters of the camera? what is the camera matrix? (intrinsic+extrinsic) General strategy: view calibration.
Camera calibration and epipolar geometry
CH24 in Robotics Handbook Presented by Wen Li Ph.D. student Texas A&M University.
Geometry of Images Pinhole camera, projection A taste of projective geometry Two view geometry:  Homography  Epipolar geometry, the essential matrix.
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Uncalibrated Geometry & Stratification Sastry and Yang
CAU Kiel DAGM 2001-Tutorial on Visual-Geometric 3-D Scene Reconstruction 1 The plan for today Leftovers and from last time Camera matrix Part A) Notation,
Lecture 12: Projection CS4670: Computer Vision Noah Snavely “The School of Athens,” Raphael.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
3D Computer Vision and Video Computing 3D Vision Lecture 15 Stereo Vision (II) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
Single-view geometry Odilon Redon, Cyclops, 1914.
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
The Pinhole Camera Model
CS223b, Jana Kosecka Rigid Body Motion and Image Formation.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
CSE473/573 – Stereo Correspondence
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Stereo Sebastian Thrun, Gary Bradski, Daniel Russakoff Stanford CS223B Computer Vision (with slides by James Rehg and.
Two-views geometry Outline Background: Camera, Projection Necessary tools: A taste of projective geometry Two view geometry:  Homography  Epipolar geometry,
Stockman MSU/CSE Math models 3D to 2D Affine transformations in 3D; Projections 3D to 2D; Derivation of camera matrix form.
Today: Calibration What are the camera parameters?
Project 4 Results Representation – SIFT and HoG are popular and successful. Data – Hugely varying results from hard mining. Learning – Non-linear classifier.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
Automatic Camera Calibration
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
Camera Geometry and Calibration Thanks to Martial Hebert.
Epipolar geometry The fundamental matrix and the tensor
Homogeneous Coordinates (Projective Space) Let be a point in Euclidean space Change to homogeneous coordinates: Defined up to scale: Can go back to non-homogeneous.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Geometric Models & Camera Calibration
ISAN-DSP GROUP Introduction to Image Processing The First Semester of Class 2546 Dr. Nawapak Eua-Anant Department of Computer Engineering Khon.
Stereo Course web page: vision.cis.udel.edu/~cv April 11, 2003  Lecture 21.
Geometric Camera Models
Cmput412 3D vision and sensing 3D modeling from images can be complex 90 horizon 3D measurements from images can be wrong.
1 Formation et Analyse d’Images Session 7 Daniela Hall 25 November 2004.
Vision Review: Image Formation Course web page: September 10, 2002.
Peripheral drift illusion. Multiple views Hartley and Zisserman Lowe stereo vision structure from motion optical flow.
Computer Vision : CISC 4/689 Going Back a little Cameras.ppt.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Single-view geometry Odilon Redon, Cyclops, 1914.
Bahadir K. Gunturk1 Phase Correlation Bahadir K. Gunturk2 Phase Correlation Take cross correlation Take inverse Fourier transform  Location of the impulse.
EECS 274 Computer Vision Geometric Camera Calibration.
stereo Outline : Remind class of 3d geometry Introduction
Feature Matching. Feature Space Outlier Rejection.
Computer vision: models, learning and inference M Ahad Multiple Cameras
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Intelligent Vision Systems Image Geometry and Acquisition ENT 496 Ms. HEMA C.R. Lecture 2.
3D Reconstruction Using Image Sequence
Camera Model Calibration
Single-view geometry Odilon Redon, Cyclops, 1914.
Digital Image Processing Additional Material : Imaging Geometry 11 September 2006 Digital Image Processing Additional Material : Imaging Geometry 11 September.
Stereo March 8, 2007 Suggested Reading: Horn Chapter 13.
Announcements Final is Thursday, March 18, 10:30-12:20 –MGH 287 Sample final out today.
Lec 26: Fundamental Matrix CS4670 / 5670: Computer Vision Kavita Bala.
CS682, Jana Kosecka Rigid Body Motion and Image Formation Jana Kosecka
Depth from disparity (x´,y´)=(x+D(x,y), y)
Two-view geometry Computer Vision Spring 2018, Lecture 10
Epipolar geometry.
Common Classification Tasks
Multiple View Geometry for Robotics
Single-view geometry Odilon Redon, Cyclops, 1914.
The Pinhole Camera Model
Presentation transcript:

Introduction to Robot Vision Ziv Yaniv Computer Aided Interventions and Medical Robotics, Georgetown University

Vision The special sense by which the qualities of an object (as color, luminosity, shape, and size) constituting its appearance are perceived through a process in which light rays entering the eye are transformed by the retina into electrical signals that are transmitted to the brain via the optic nerve. [Miriam Webster dictionary]

Single Lens Reflex (SLR) Camera The Sensor endoscope Single Lens Reflex (SLR) Camera webcam C-arm X-ray

The Sensor Model: Pin-hole Camera, Perspective Projection y z x Image plane optical axis focal point x y z principle point

Machine Vision Goal: Model: Obtain useful information about the 3D world from 2D images. Model: Regions Textures Corners Lines … 3D Geometry Object identification Activity detection … images actions

Machine Vision Goal: High level (machine learning/pattern recognition) Obtain useful information about the 3D world from 2D images. Low level (image processing) image filtering (smoothing, histogram modification…), feature extraction (corner detection, edge detection,…) stereo vision shape from X (shading, motion,…) … High level (machine learning/pattern recognition) object detection object recognition clustering

Machine Vision How hard can it be?

Machine Vision How hard can it be?

Robot Vision Simultaneous Localization and Mapping (SLAM) Visual Servoing.

Robot Vision Simultaneous Localization and Mapping (SLAM) – create a 3D map of the world and localize within this map. NASA stereo vision image processing, as used by the MER Mars rovers

Robot Vision Simultaneous Localization and Mapping (SLAM) – create a 3D map of the world and localize within this map. “Simultaneous Localization and Mapping with Active Stereo Vision”, J. Diebel, K. Reuterswärd, S. Thrun, J. Davis, R. Gupta, IROS 2004.

Robot Vision Visual Servoing – Using visual feedback to control a robot: image-based systems: desired motion directly from image. “An image-based visual servoing scheme for following paths with nonholonomic mobile robots” A. Cherubini, F. Chaumette, G. Oriolo, ICARCV 2008.

Robot Vision Visual Servoing – Using visual feedback to control a robot: Position-based systems: desired motion from 3D reconstruction estimated from image.

System Configuration Difficulty of similar tasks in different settings varies widely: How many cameras? Are the cameras calibrated? What is the camera-robot configuration? Is the system calibrated (hand-eye calibration)? Common configurations: x y z x y z x y z x y z x y z x y z x y z

System Characteristics The greater the control over the system configuration and environment the easier it is to execute a task. System accuracy is directly dependent upon model accuracy – what accuracy does the task require?. All measurements and derived quantitative values have an associated error.

Stereo Reconstruction Compute the 3D location of a point in the stereo rig’s coordinate system: Rigid transformation between the two cameras is known. Cameras are calibrated –given a point in the world coordinate system we know how to map it to the image. Same point localized in the two images. Camera 1 Camera 2 world

Commercial Stereo Vision Polaris Vicra infra-red system (Northern Digitial Inc.) MicronTracker visible light system (Claron Technology Inc.)

Commercial Stereo Vision Images acquired by the Polaris Vicra infra-red stereo system: left image right image

Stereo Reconstruction Wide or short baseline – reconstruction accuracy vs. difficulty of point matching Camera 1 Camera 2 Camera 1 Camera 2 Camera 1 Camera 2

Camera Model Points P, p, and O, given in the camera coordinate system, are collinear. There is a number a for which O + aP = p aP = p a = f/Z , therefore x y z P=[X,Y,Z] O p=[x,y,f] f

Camera Model Transform the pixel coordinates from the camera coordinate system to the image coordinate system: Image origin (principle point) is at [x0,y0] relative to the camera coordinate system. Need to change from metric units to pixels, scaling factors kx, ky. x y principle point [x’,y’] Finally, the image coordinate system may be skewed resulting in:

Camera Model As our original assumption was that points are given in the camera coordinate system, a complete projection matrix is of the form: C – camera origin in the world coordinate system. How many degrees of freedom does M have?

world coordinate system Camera Calibration Given pairs of points, piT=[x,y,w], PiT=[X,Y,Z,W], in homogenous coordinates we have: image coordinate system z x calibration object/ world coordinate system y principle point Our goal is to estimate M y z x camera coordinate system As the points are in homogenous coordinates the vectors p and MP are not necessarily equal, they have the same direction but may differ by a non-zero scale factor.

Camera Calibration After a bit of algebra we have: The three equations are linearly dependent: Each point pair contributes two equations. Exact solution: M has 11 degrees of freedom, requiring a minimum of n=6 pairs. Least squares solution: For n>6 minimize ||Am|| s.t. ||m||=1.

Obtaining the Rays Camera location in the calibration object’s coordinate system, C, is given by the one dimensional right null space of the matrix M (MC=0). A 3D homogenous point P = M+p is on the ray defined by p and the camera center [it projects onto p, MM+p =Ip=p]. These two points define our ray in the world coordinate system. As both cameras were calibrated with respect to the same coordinate system the rays will be in the same system too.

Intersecting the Rays

World vs. Model Actual cameras most often don’t follow the ideal pin-hole model, usually exhibit some form of distortion (barrel, pin-cushion, S). Sometimes the world changes to fit your model, improvements in camera/lens quality can improve model performance. old image-Intensifier x-ray: pin-hole+distortion replaced by flat panel x-ray: pin-hole

Additional Material Code: Machine Vision: Camera calibration toolbox for matlab (Jean-Yves Bouguet ) http://www.vision.caltech.edu/bouguetj/calib_doc/ Machine Vision: “Multiple View Geometry in Computer Vision”, Hartley and Zisserman, Cambridge University Press. "Machine Vision", Jain, Kasturi, Schunck, McGraw-Hill. Robot Vision: “Simultaneous Localization and Mapping: Part I”, H. Durant-Whyte, T. Bailey, IEEE Robotics and Automation Magazine, Vol. 13(2), pp. 99-110, 2006. “Simultaneous Localization and Mapping (SLAM) : Part II”,T. Bailey, H. Durant-Whyte, IEEE Robotics and Automation Magazine, Vol. 13(3), pp. 108-117, 2006. “Visual Servo Control Part I: Basic Approaches”, IEEE Robotics and Automation Magazine, Vol. 13(4), 82-90, 2006. Visual Servo Control Part II: Advanced Approaches”, IEEE Robotics and Automation Magazine, Vol. 14(1), 109-118, 2007.