Announcements Final Exam May 13th, 8 am (not my idea).

Slides:



Advertisements
Similar presentations
Object Recognition from Local Scale-Invariant Features David G. Lowe Presented by Ashley L. Kapron.
Advertisements

Feature Based Image Mosaicing
Image Registration  Mapping of Evolution. Registration Goals Assume the correspondences are known Find such f() and g() such that the images are best.
Alignment Visual Recognition “Straighten your paths” Isaiah.
QR Code Recognition Based On Image Processing
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Last 4 lectures Camera Structure HDR Image Filtering Image Transform.
Document Image Processing
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Computer Vision - A Modern Approach Set: Model-based Vision Slides by D.A. Forsyth Recognition by Hypothesize and Test General idea –Hypothesize object.
University of Pennsylvania 1 GRASP CIS 580 Machine Perception Fall 2004 Jianbo Shi Object recognition.
Computer Vision Detecting the existence, pose and position of known objects within an image Michael Horne, Philip Sterne (Supervisor)
Segmentation (2): edge detection
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Image alignment Image from
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
November 4, 2014Computer Vision Lecture 15: Shape Representation II 1Signature Another popular method of representing shape is called the signature. In.
Computer Vision Optical Flow
Image alignment Image from
Lecture 5 Template matching
Announcements Final Exam May 16 th, 8 am (not my idea). Practice quiz handout 5/8. Review session: think about good times. PS5: For challenge problems,
Announcements Quiz Thursday Quiz Review Tomorrow: AV Williams 4424, 4pm. Practice Quiz handout.
Problem Sets Problem Set 3 –Distributed Tuesday, 3/18. –Due Thursday, 4/3 Problem Set 4 –Distributed Tuesday, 4/1 –Due Tuesday, 4/15. Probably a total.
Chamfer Matching & Hausdorff Distance Presented by Ankur Datta Slides Courtesy Mark Bouts Arasanathan Thayananthan.
Announcements Take home quiz given out Thursday 10/23 –Due 10/30.
Scale Invariant Feature Transform (SIFT)
Recognition by Linear Combinations of Models By Shimon Ullman & Rosen Basri Presented by: Piotr Dollar Nov. 19, 2002.
1Ellen L. Walker Matching Find a smaller image in a larger image Applications Find object / pattern of interest in a larger picture Identify moving objects.
MSU CSE 803 Stockman CV: Matching in 2D Matching 2D images to 2D images; Matching 2D images to 2D maps or 2D models; Matching 2D maps to 2D maps.
May 2004Stereo1 Introduction to Computer Vision CS / ECE 181B Tuesday, May 11, 2004  Multiple view geometry and stereo  Handout #6 available (check with.
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Robust estimation Problem: we want to determine the displacement (u,v) between pairs of images. We are given 100 points with a correlation score computed.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Computing transformations Prof. Noah Snavely CS1114
כמה מהתעשייה? מבנה הקורס השתנה Computer vision.
Robust fitting Prof. Noah Snavely CS1114
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean Hall 5409 T-R 10:30am – 11:50am.
By Yevgeny Yusepovsky & Diana Tsamalashvili the supervisor: Arie Nakhmani 08/07/2010 1Control and Robotics Labaratory.
October 14, 2014Computer Vision Lecture 11: Image Segmentation I 1Contours How should we represent contours? A good contour representation should meet.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
A 3D Model Alignment and Retrieval System Ding-Yun Chen and Ming Ouhyoung.
CSCE 643 Computer Vision: Structure from Motion
Digital Image Processing CCS331 Relationships of Pixel 1.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
CS654: Digital Image Analysis Lecture 25: Hough Transform Slide credits: Guillermo Sapiro, Mubarak Shah, Derek Hoiem.
21 June 2009Robust Feature Matching in 2.3μs1 Simon Taylor Edward Rosten Tom Drummond University of Cambridge.
EECS 274 Computer Vision Model Fitting. Fitting Choose a parametric object/some objects to represent a set of points Three main questions: –what object.
Computer Vision Set: Object Recognition Slides by C.F. Olson 1 Object Recognition.
CSE 185 Introduction to Computer Vision Feature Matching.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
MultiModality Registration Using Hilbert-Schmidt Estimators By: Srinivas Peddi Computer Integrated Surgery II April 6 th, 2001.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Dynamic Programming (DP), Shortest Paths (SP)
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Hough Transform CS 691 E Spring Outline Hough transform Homography Reading: FP Chapter 15.1 (text) Some slides from Lazebnik.
SIFT.
776 Computer Vision Jan-Michael Frahm Spring 2012.
SIFT Scale-Invariant Feature Transform David Lowe
Nearest-neighbor matching to feature database
A special case of calibration
Object recognition Prof. Graeme Bailey
Nearest-neighbor matching to feature database
Geometric Hashing: An Overview
CV: Matching in 2D Matching 2D images to 2D images; Matching 2D images to 2D maps or 2D models; Matching 2D maps to 2D maps MSU CSE 803 Stockman.
SIFT.
CSE 185 Introduction to Computer Vision
Fourier Transform of Boundaries
Image Registration  Mapping of Evolution
Presentation transcript:

Announcements Final Exam May 13th, 8 am (not my idea). Review session: 11am Wed. AVW 4424 Questions on problem set?

Recognizing Objects: Feature Matching Problem: Match when viewing conditions change a lot. Lighting changes: brightness constancy false. Viewpoint changes: appearance changes, many viewpoints. One Solution: Match using edge-based features. Edges less variable to lighting, viewpoint. More compact representation can lead to efficiency. Match image or object to image If object, matching may be asymmetric Object may be 3D. Line finding was an example: line=object; points=image.

Problem Definition An Image is a set of 2D geometric features, along with positions. An Object is a set of 2D/3D geometric features, along with positions. A pose positions the object relative to the image. 2D Translation; 2D translation + rotation; 2D translation, rotation and scale; planar or 3D object positioned in 3D with perspective or scaled orth. The best pose places the object features nearest the image features

Strategy Build feature descriptions Search possible poses. Can search space of poses Or search feature matches, which produce pose Transform model by pose. Compare transformed model and image. Pick best pose.

Presentation Strategy Already discussed finding features. First discuss picking best pose since this defines the problem. Second discuss search methods appropriate for 2D. Third discuss transforming model in 2D and 3D. Fourth discuss search for 3D objects.

Example

Evaluate Pose We look at this first, since it defines the problem. Again, no perfect measure; Trade-offs between veracity of measure and computational considerations.

Chamfer Matching For every edge point in the transformed object, compute the distance to the nearest image edge point. Sum distances.

Main Feature: Every model point matches an image point. An image point can match 0, 1, or more model points.

Variations Sum a different distance f(d) = d2 or Manhattan distance. f(d) = 1 if d > threshold, 0 otherwise. This is called bounded error. Use maximum distance instead of sum. This is called: directed Hausdorff distance. Use other features Corners. Lines. Then position and angles of lines must be similar. Model line may be subset of image line.

Other comparisons Enforce each image feature can match only one model feature. Enforce continuity, ordering along curves. These are more complex to optimize.

Presentation Strategy Already discussed finding features. First discuss picking best pose since this defines the problem. Second discuss search methods appropriate for 2D. Third discuss transforming model in 2D and 3D. Fourth discuss search for 3D objects.

Pose Search Simplest approach is to try every pose. Two problems: many poses, costly to evaluate each. We can reduce the second problem with:

Pose: Chamfer Matching with the Distance Transform 2 1 1 2 3 2 1 1 2 3 2 1 2 1 3 2 1 1 2 3 2 1 1 2 3 3 2 1 1 3 4 3 2 1 1 Example: Each pixel has (Manhattan) distance to nearest edge pixel.

D.T. Adds Efficiency Compute once. Fast algorithms to compute it. Makes Chamfer Matching simple.

Then, try all translations of model edges Then, try all translations of model edges. Add distances under each edge pixel. 2 1 1 2 3 2 1 1 2 3 2 1 2 1 3 2 1 1 2 3 2 1 1 2 3 3 2 1 1 3 4 3 2 1 1

Computing Distance Transform It’s only done once, per problem, not once per pose. Basically a shortest path problem. Simple solution passing through image once for each distance. First pass mark edges 0. Second, mark 1 anything next to 0, unless it’s already marked. Etc…. Actually, a more clever method requires 2 passes.

Announcements Practice problems for final here and on web page. Please complete on-line evaluations: https://www.courses.umd.edu/online_evaluation/ Final Thursday, May 13, 8am. Review Wed., May 12, 11am, AV Williams 4424. Hand in Problem Set 7

Pose: Ransac Match enough features in model to features in image to determine pose. Examples: match a point and determine translation. match a corner and determine translation and rotation. Points and translation, rotation, scaling? Lines and rotation and translation?

RANSAC Matching k model and image features determines pose. Match k model and image features at Random. Determine model pose. Project remaining model features into image. Count number of model features within e pixels of an image feature. Repeat. Pick pose that matches most features.

Issues for RANSAC How do we determine pose? How many matches do we need to try?

Pose: Generalized Hough Transform Like Hough Transform, but for general shapes. Example: match one point to one point, and for every rotation of the object its translation is determined.

Presentation Strategy Already discussed finding features. First discuss picking best pose since this defines the problem. Second discuss search methods appropriate for 2D. Third discuss transforming model in 2D and 3D. Fourth discuss search for 3D objects.

Computing Pose: Points Solve I = S*P. In Structure-from-Motion, we knew I. In Recognition, we know P. This is just set of linear equations Ok, maybe with some non-linear constraints.

Linear Pose: 2D Translation We know x,y,u,v, need to find translation. For one point, u1 - x1 = tx ; v1 - x1 = ty For more points we can solve a least squares problem.

Linear Pose: 2D rotation, translation and scale Notice a and b can take on any values. Equations linear in a, b, translation. Solve exactly with 2 points, or overconstrained system with more.

Linear Pose: 3D Affine

Pose: Scaled Orthographic Projection of Planar points s1,3, s2,3 disappear. Non-linear constraints disappear with them.

Non-linear pose A bit trickier. Some results: 2D rotation and translation. Need 2 points. 3D scaled orthographic. Need 3 points, give 2 solutions. 3D perspective, camera known. Need 3 points. Solve 4th degree polynomial. 4 solutions.

Transforming the Object We don’t really want to know pose, we want to know what the object looks like in that pose. We start with: Solve for pose: Project rest of points:

Transforming object with Linear Combinations No 3D model, but we’ve seen object twice before. See four points in third image, need to fill in location of other points. Just use rank theorem.

Recap: Recognition w/ RANSAC Find features in model and image. Such as corners. Match enough to determine pose. Such as 3 points for planar object, scaled orthographic projection. Determine pose. Project rest of object features into image. Look to see how many image features they match. Example: with bounded error, count how many object features project near an image feature. Repeat steps 2-5 a bunch of times. Pick pose that matches most features.

Figure 18.1 - caption tells the story Figure from “Object recognition using alignment,” D.P. Huttenlocher and S. Ullman, Proc. Int. Conf. Computer Vision, 1986, copyright IEEE, 1986

Recognizing 3D Objects Previous approach will work. But slow. RANSAC considers n3m3 possible matches. About m3 correct. Solutions: Grouping. Find features coming from single object. Viewpoint invariance. Match to small set of model features that could produce them.

Grouping: Continuity

Connectivity Connected lines likely to come from boundary of same object. Boundary of object often produces connected contours. Different objects more rarely do; only when overlapping. Connected image lines match connected model lines. Disconnected model lines generally don’t appear connected.

Other Viewpoint Invariants Parallelism Convexity Common region ….

Planar Invariants A t

p3 p4 p1 p2 p4 = p1 + a(p2-p1) + b(p3-p1) A(p4)+ t = A(p1+a(p2-p1) + b(p3-p1)) + t = A(p1)+t + a(A(p2)+t – A(p1)-t) + b(A(p3)+t – A(p1)-t) p4 is linear combination of p1,p2,p3. Transformed p4 is same linear combination of transformed p1, p2, p3.

Here segmentation is being used to break up a single MRI slice into regions Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

which are assembled into a 3D model like so Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

Patient is being imaged with a scanner Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

A view showing the patient and model details superimposed Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

A patient with a view of the tissue the surgeon needs to remove superimposed; the view is registered to the patient, so its “in the right place” Figures by kind permission of Eric Grimson; further information can be obtained from his web site http://www.ai.mit.edu/people/welg/welg.html.

What we didn’t talk about Smooth 3D objects. Can we find the guaranteed optimal solution? Indexing with invariants. Error propagation.