Heather Dunlop : Advanced Perception January 25, 2006

Slides:



Advertisements
Similar presentations
OpenCV Introduction Hang Xiao Oct 26, History  1999 Jan : lanched by Intel, real time machine vision library for UI, optimized code for intel 
Advertisements

Boundary Detection - Edges Boundaries of objects –Usually different materials/orientations, intensity changes.
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Computer Vision Lecture 16: Texture
Chapter 4: Linear Models for Classification
ADS lab NCKU1 Michael Maire, Pablo Arbelaez, Charless Fowlkes, and Jitendra Malik university of California, Berkeley – Berkeley university of California,
1 P. Arbelaez, M. Maire, C. Fowlkes, J. Malik. Contour Detection and Hierarchical image Segmentation. IEEE Trans. on PAMI, Student: Hsin-Min Cheng.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Boundary Extraction in Natural Images Using Ultrametric Contour Maps Pablo Arbeláez Université Paris Dauphine Presented by Derek Hoiem.
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
1Ellen L. Walker Edges Humans easily understand “line drawings” as pictures.
Fast intersection kernel SVMs for Realtime Object Detection
EE663 Image Processing Edge Detection 1
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
1 Texture Texture is a description of the spatial arrangement of color or intensities in an image or a selected region of an image. Structural approach:
Learning to Detect Natural Image Boundaries Using Local Brightness, Color, and Texture Cues David R. Martin Charless C. Fowlkes Jitendra Malik.
Edge detection. Edge Detection in Images Finding the contour of objects in a scene.
EE663 Image Processing Edge Detection 2 Dr. Samir H. Abdul-Jauwad Electrical Engineering Department King Fahd University of Petroleum & Minerals.
Texture Texture is a description of the spatial arrangement of color or intensities in an image or a selected region of an image. Structural approach:
Abstract We present a model of curvilinear grouping using piecewise linear representations of contours and a conditional random field to capture continuity.
Measuring the Ecological Statistics of Figure-Ground Charless Fowlkes, David Martin, Jitendra Malik.
1 Learning to Detect Natural Image Boundaries David Martin, Charless Fowlkes, Jitendra Malik Computer Science Division University of California at Berkeley.
Berkeley Vision GroupNIPS Vancouver Learning to Detect Natural Image Boundaries Using Local Brightness,
Canny Edge Detector1 1)Smooth image with a Gaussian optimizes the trade-off between noise filtering and edge localization 2)Compute the Gradient magnitude.
Measuring the Ecological Statistics of Figure-Ground Charless Fowlkes, David Martin, Jitendra Malik.
A Database of Human Segmented Natural Images and Two Applications David Martin, Charless Fowlkes, Doron Tal, Jitendra Malik UC Berkeley
1 The Ecological Statistics of Grouping by Similarity Charless Fowlkes, David Martin, Jitendra Malik Computer Science Division University of California.
Color a* b* Brightness L* Texture Original Image Features Feature combination E D 22 Boundary Processing Textons A B C A B C 22 Region Processing.
Texture Readings: Ch 7: all of it plus Carson paper
Filters and Edges. Zebra convolved with Leopard.
Lecture 2: Image filtering
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
1 Ecological Statistics and Perceptual Organization Charless Fowlkes work with David Martin and Jitendra Malik at University of California at Berkeley.
The blue and green colors are actually the same.
1 Image Features Hao Jiang Sept Image Matching 2.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Overview Introduction to local features
Overview Harris interest points Comparing interest points (SSD, ZNCC, SIFT) Scale & affine invariant interest points Evaluation and comparison of different.
1 Contours and Junctions in Natural Images Jitendra Malik University of California at Berkeley (with Jianbo Shi, Thomas Leung, Serge Belongie, Charless.
Lecture 6: Edge Detection CAP 5415: Computer Vision Fall 2008.
Edge detection Goal: Identify sudden changes (discontinuities) in an image Intuitively, most semantic and shape information from the image can be encoded.
CSE 185 Introduction to Computer Vision Edges. Scale space Reading: Chapter 3 of S.
Ilya Gurvich 1 An Undergraduate Project under the supervision of Dr. Tammy Avraham Conducted at the ISL Lab, CS, Technion.
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
Kylie Gorman WEEK 1-2 REVIEW. CONVERTING AN IMAGE FROM RGB TO HSV AND DISPLAY CHANNELS.
Chapter 9: Image Segmentation
A New Method for Crater Detection Heather Dunlop November 2, 2006.
A Statistical Approach to Texture Classification Nicholas Chan Heather Dunlop Project Dec. 14, 2005.
Segmentation Through Optimization Pyry Matikainen.
Lecture 04 Edge Detection Lecture 04 Edge Detection Mata kuliah: T Computer Vision Tahun: 2010.
Digital Image Processing Lecture 17: Segmentation: Canny Edge Detector & Hough Transform Prof. Charlene Tsai.
Edge Segmentation in Computer Images CSE350/ Sep 03.
Computer Vision Image Features Instructor: Dr. Sherif Sami Lecture 4.
Image Features (I) Dr. Chang Shu COMP 4900C Winter 2008.
Finding Boundaries Computer Vision CS 143, Brown James Hays 09/28/11 Many slides from Lana Lazebnik, Steve Seitz, David Forsyth, David Lowe, Fei-Fei Li,
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
SIFT.
Edge Detection Images and slides from: James Hayes, Brown University, Computer Vision course Svetlana Lazebnik, University of North Carolina at Chapel.
SIFT Scale-Invariant Feature Transform David Lowe
CS262: Computer Vision Lect 09: SIFT Descriptors
Jeremy Bolton, PhD Assistant Teaching Professor
ECE 692 – Advanced Topics in Computer Vision
Computer Vision Lecture 16: Texture II
Presented by: Yang Yu Spatiotemporal GMM for Background Subtraction with Superpixel Hierarchy Mingliang Chen, Xing Wei, Qingxiong.
Texture.
Evolving Logical-Linear Edge Detector with Evolutionary Algorithms
Grouping/Segmentation
Blobworld Texture Features
Presentation transcript:

Heather Dunlop 16-721: Advanced Perception January 25, 2006 Learning to Detect Natural Image Boundaries Using Local Brightness, Color and Texture Cues by David R. Martin, Charless C. Fowlkes, Jitendra Malik Heather Dunlop 16-721: Advanced Perception January 25, 2006

What is a Boundary? Canny Human Martin, 2002 “A boundary is a contour in the image plane that represents a change in pixel ownership from one object or surface to another” Edges are not boundaries

Dataset “You will be presented a photographic image. Divide the image into some number of segments, where the segments represent ‘things’ or ‘parts of things’ in the scene. The number of segments is up to you, as it depends on the image. Something between 2 and 30 is likely to be appropriate. It is important that all of the segments have approximately equal importance.”

Dataset Database of over 1000 images and 5-10 segmentations for each Thicker lines indicate more common boundary choice Martin, 2002

Boundaries Non-boundaries Boundaries Intensity Brightness Color Brightness on its own is not sufficient Color Texture Martin, 2002

Method Goal: learn the probability of a boundary, Pb(x,y,θ) Image Optimized Cues Boundary Strength Brightness Color Texture Benchmark Human Segmentations Cue Combination Model Goal: “use features extracted from such an image patch to estimate the posterior probability of a boundary passing through the center point” Use cues such as intensity, brightness, color and texture to get a measure of boundary strength How to combine cues? It’s a supervised learning problem. Learn an optimal local boundary model from labeled images Approach: look at each pixel for local discontinuities in several feature channels, over a range of orientations and scales Martin, 2002

Image Features CIE L*a*b* color space (luminance, red-green, yellow-blue) Oriented Energy: fe: Gaussian second derivative fo: Its Hilbert transform Brightness L* distribution Color a* and b* distributions (joint or marginal) Texture “In natural images, brightness edges are more than simple steps. Phenomena such as specularities, mutual illumination, and shading result in composite intensity profiles consisting of steps, peaks and roofs.” even and odd symmetric filters has maximum response for contours at orientation theta -minimal accuracy difference between joint and marginal distributions -joint is much more computation intensive, so use marginal

Texture Convolve with a filter bank: Gaussian second derivative Its Hilbert transform Difference of Gaussians Filter responses give a measure of texture Each pixel is associated with a vector of 13 filter responses centered at that pixel

Other Filter Banks Leung-Malik filter set: Schmid filter set: Maximum Response 8 filter set: MR8: take maximum over orientations

Textons Convolve image with filter bank Cluster filter responses to form textons Adapted from Martin, 2002 and Varma, Zisserman, 2005

Texton Distribution Assign each pixel to nearest texton Form distribution of textons Adapted from Martin, 2002 and Varma, Zisserman, 2005

Gradient-based Features Brightness (BG), color (CG), texture (TG) gradients Half-disc regions described by histograms Compare distributions with χ2 statistic  r (x,y) “At at location (x,y) in the image, draw a circle of radius r, and divide it along the diameter at orientation theta. The gradient function G(x,y,theta,r) compares the contents of the two resulting disc halves. A large difference between the disc halves indicates a discontinuity in the image along the disc’s diameter.” 8 orientations, 3 scales

Texture Gradient Texton distribution in two half circles Martin, 2002

Localization Tightly localize boundaries Reduce noise Coalesce double detections Improve OE and TG features OE OE localized Fit a peak (parabola) TG TG localized Martin, Fowlkes, Malik, 2004

Optimization Texture parameters: type of filter bank scale of filters number of textons universal or image-specific textons Other possible distance/histogram comparison metrics Number of bins for histograms Scale parameter for all cues It was found that a single scale is sufficient for texture Universal vs. image-specific textons: -computational cost approximately equal -accuracy approx equal -optimal number of textons for universal textons is roughly double that for image-specific textons -universal preferable if doing image retrieval and object recognition -image-specific used for convenience L1, L2 norm, chi-square, quadratic form, Earth Mover’s distance

Evaluation Methodology Posterior probability of boundary: Pb(x,y,θ) Evaluation measure: precision recall curve F-measure: “formulate boundary-detection as a classification problem of discriminating non-boundary from boundary pixels” -take maximum over orientations PR curve: “captures the tradeoff between accuracy and noise as the detector threshold varies” “Precision is the fraction of detections that are true positives rather than false positives, while recall is the fraction of true positives that are detected rather than missed. In probabilistic terms, precision is the probability that the detector’s signal is valid, and recall is the probability that the ground truth data was detected.” “Each point on the PR curve is computed from the detector’s output at a particular threshold.” “First, we correspond the machine boundary map separately with each human map in turn. Only those machine boundary pixels that match no human boundary are counted as false positives. The hit rate is simply averaged over the different human, so that to achieve perfect recall the machine boundary map must explain all of the human data.” F-measure: defines a relative cost between precision and recall for a specific application The location of maximum F-measure along the curve provides the optimal detector threshold given alpha. Canny’s goals of boundary detection: high detection rate, single detection, good localization Martin, 2002

Cue Combination Which cues should be used? OE is redundant when other cues are present BG+CG+TG produces best results Martin, 2002

Classifiers Until now, only logistic regression was used Other possible classifiers: Density estimation Classification trees Hierarchical mixtures of experts Support vector machines Logistic regression: fast convergence and reliable svm: prohibitively slow didn’t always produce meaningful results performance of all classifiers is approximately equal Martin, 2002

Result Comparison Alternative methods: Matlab’s Canny edge detector with and without hysteresis Spatially-averaged second moment matrix (2MM) Where’s the human curve come from? “The points marked by a ‘+’ on the plot show the precision and recall of each ground truth human segmentation when compared to other humans. . . . The solid curve shows the F=0.80 curve, representing the frontier of human performance for this task.” Martin, 2002

Results Canny 2MM BG+CG+TG Human Image Martin, 2002 -texture on man’s shirt -no boundary on shoulder Martin, 2002

Results Canny 2MM BG+CG+TG Human Image Martin, 2002 -windows are texture not boundary -underside of boat and on dock Martin, 2002

Results Canny 2MM BG+CG+TG Human Image Martin, 2002

Conclusions Large data set used for testing Texture gradients are a powerful cue Simple linear model sufficient for cue combination Outperforms existing methods An approach that is useful for higher-level algorithms Code is available online: http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbench/