Learning to Detect Natural Image Boundaries Using Local Brightness, Color, and Texture Cues David R. Martin Charless C. Fowlkes Jitendra Malik.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pose Estimation and Segmentation of People in 3D Movies Karteek Alahari, Guillaume Seguin, Josef Sivic, Ivan Laptev Inria, Ecole Normale Superieure ICCV.
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Classification using intersection kernel SVMs is efficient Joint work with Subhransu Maji and Alex Berg Jitendra Malik UC Berkeley.
ADS lab NCKU1 Michael Maire, Pablo Arbelaez, Charless Fowlkes, and Jitendra Malik university of California, Berkeley – Berkeley university of California,
1 P. Arbelaez, M. Maire, C. Fowlkes, J. Malik. Contour Detection and Hierarchical image Segmentation. IEEE Trans. on PAMI, Student: Hsin-Min Cheng.
Boundary Extraction in Natural Images Using Ultrametric Contour Maps Pablo Arbeláez Université Paris Dauphine Presented by Derek Hoiem.
Ghunhui Gu, Joseph J. Lim, Pablo Arbeláez, Jitendra Malik University of California at Berkeley Berkeley, CA
Fast intersection kernel SVMs for Realtime Object Detection
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Segmentation Divide the image into segments. Each segment:
Abstract We present a model of curvilinear grouping using piecewise linear representations of contours and a conditional random field to capture continuity.
Measuring the Ecological Statistics of Figure-Ground Charless Fowlkes, David Martin, Jitendra Malik.
1 Learning to Detect Natural Image Boundaries David Martin, Charless Fowlkes, Jitendra Malik Computer Science Division University of California at Berkeley.
CVR05 University of California Berkeley 1 Familiar Configuration Enables Figure/Ground Assignment in Natural Scenes Xiaofeng Ren, Charless Fowlkes, Jitendra.
CS223B Assignment 1 Recap. Lots of Solutions! 37 Groups Many different approaches Let’s take a peek at all 37 results on one image from the test set.
Berkeley Vision GroupNIPS Vancouver Learning to Detect Natural Image Boundaries Using Local Brightness,
Measuring the Ecological Statistics of Figure-Ground Charless Fowlkes, David Martin, Jitendra Malik.
Segmentation by Clustering Reading: Chapter 14 (skip 14.5) Data reduction - obtain a compact representation for interesting image data in terms of a set.
1 The Ecological Statistics of Grouping by Similarity Charless Fowlkes, David Martin, Jitendra Malik Computer Science Division University of California.
Color a* b* Brightness L* Texture Original Image Features Feature combination E D 22 Boundary Processing Textons A B C A B C 22 Region Processing.
1 How do ideas from perceptual organization relate to natural scenes?
1 Ecological Statistics and Perceptual Organization Charless Fowlkes work with David Martin and Jitendra Malik at University of California at Berkeley.
The blue and green colors are actually the same.
Cue Integration in Figure/Ground Labeling Xiaofeng Ren, Charless Fowlkes and Jitendra Malik, U.C. Berkeley We present a model of edge and region grouping.
Heather Dunlop : Advanced Perception January 25, 2006
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Entropy and some applications in image processing Neucimar J. Leite Institute of Computing
Segmentation and Grouping Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/23/10.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.
Reporter: Fei-Fei Chen. Wide-baseline matching Object recognition Texture recognition Scene classification Robot wandering Motion tracking.
Recognition using Regions (Demo) Sudheendra V. Outline Generating multiple segmentations –Normalized cuts [Ren & Malik (2003)] Uniform regions –Watershed.
1 Contours and Junctions in Natural Images Jitendra Malik University of California at Berkeley (with Jianbo Shi, Thomas Leung, Serge Belongie, Charless.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
CSE 185 Introduction to Computer Vision Pattern Recognition 2.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
EDGE DETECTION IN COMPUTER VISION SYSTEMS PRESENTATION BY : ATUL CHOPRA JUNE EE-6358 COMPUTER VISION UNIVERSITY OF TEXAS AT ARLINGTON.
Phase Congruency Detects Corners and Edges Peter Kovesi School of Computer Science & Software Engineering The University of Western Australia.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
CS654: Digital Image Analysis Lecture 30: Clustering based Segmentation Slides are adapted from:
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
CSE 185 Introduction to Computer Vision Edges. Scale space Reading: Chapter 3 of S.
Ilya Gurvich 1 An Undergraduate Project under the supervision of Dr. Tammy Avraham Conducted at the ISL Lab, CS, Technion.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe Presented by Tony X. Han March 11, 2008.
Probability and Statistics in Vision. Probability Objects not all the sameObjects not all the same – Many possible shapes for people, cars, … – Skin has.
CSE 185 Introduction to Computer Vision Feature Matching.
A New Method for Crater Detection Heather Dunlop November 2, 2006.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #23.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
Lecture 30: Segmentation CS4670 / 5670: Computer Vision Noah Snavely From Sandlot ScienceSandlot Science.
Finding Boundaries Computer Vision CS 143, Brown James Hays 09/28/11 Many slides from Lana Lazebnik, Steve Seitz, David Forsyth, David Lowe, Fei-Fei Li,
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
SIFT.
Edge Detection Images and slides from: James Hayes, Brown University, Computer Vision course Svetlana Lazebnik, University of North Carolina at Chapel.
Edge Detection slides taken and adapted from public websites:
SIFT Scale-Invariant Feature Transform David Lowe
Distinctive Image Features from Scale-Invariant Keypoints
A Tutorial on HOG Human Detection
Jeremy Bolton, PhD Assistant Teaching Professor
SIFT.
Grouping/Segmentation
Feature descriptors and matching
EM Algorithm and its Applications
Presentation transcript:

Learning to Detect Natural Image Boundaries Using Local Brightness, Color, and Texture Cues David R. Martin Charless C. Fowlkes Jitendra Malik

Paper Contribution Proposed Boundary Classifier –Uses local image features –Detect & Localize Boundaries Probability each pixel belongs to a boundary Determine boundary angle –Robust on natural scenes –Supervised-learning using human-segmented images

Feature Hierarchy Cues for perceptual grouping: –Low-Level: brightness, color, texture, depth, motion –Mid-Level: continuity, closure, convexity, symmetry, … –High-Level: familiar objects and configurations This paper –Strictly uses low-level features Comparison to other low-level classifiers –Canny –Edge detector based on Eigenvalues of “Second Moment Matrix”

Boundary Detection Other methods detect edges –Abrupt change in low-level feature –Ad-hoc additions introduce errors This method detects boundaries –Contour separating objects –Resembling human-abilities

Goal & Outline Goal: Model the posterior probability of a boundary P b (x,y,  ) at each pixel and orientation using local cues Method: Supervised learning using dataset of 12,000 segmentations of 1,000 images by 30 subjects Outline: –3 Image-feature cues: brightness, color, texture –Cue calibration –Cue combination –Compare with other approaches

Image Features Look at region around each pixel for feature discontinuities –Range of orientations –Range of scales Features –Oriented Energy (not used) –Brightness gradient –Color Gradient –Texture Gradient

Image Features All 3 Features are gradient-based Gradient found by: –For every pixel (x,y) draw circle of radius ‘r’ –Split circle in half –Compare contents of two halves G(x,y, ,r) Large difference indicates an edge –Repeat for every ‘  ’ & ‘r’ Implementation –8 orientations of ‘  ’ –3 scales ‘r’  r (x,y)

Non-BoundariesBoundaries IT BC Image Features

Brightness Features Brightness Gradient BG(x,y,r,  ) –Model intensity values in each disk-half using kernel density estimation –Create a histogram of distribution –Compare disk-halves by comparing histograms  2 difference in L * distribution

Color Features Color Gradient CG(x,y,r,  ) –Model color values in each disk-half using kernel density estimation –Color Space Red-Green (a) Yellow-Blue (b) –2D space (a*b) greatly increases computation –Instead, use two 1D spaces (a+b) –Create 2 histograms of kernel densities –Compare disk-halves by comparing histograms  2 difference in (a * distribution) + (b * distribution)

Texture Features Texture Gradient BG(x,y,r,  ) –Pixel-texture values are computed using 13 filters –Pixels are thus represented with a 13-feature vector –Each disk-half is modeled by a point cloud of vectors in 13-dimensional space –Problem: How does one compare two 13D spaces?

Texture Features Solution: Textons estimate joint distribution using adaptive bins –Filter response vectors are clustered using k-means –Cluster centers represent texture primitives (textons) –Example Texton set K = 64 Trained using 200 images

Texture Features Texture Gradient BG(x,y,r,  ) –Pixels in each disk-half are assigned to nearest texton –Disk-halfs are represented by a histogram of textons –Compare disk-halves by comparing histograms  2 difference in T * distribution Texton Map

Feature Localization Problem: Can’t localize because features don’t form sharp peaks around boundaries –Smooth peaks –Double peaks Solution: For each pixel –Use least-squares to fit a cylindrical parabola over the 2D window of radius ‘r’ –Parabola center = localized edge

Evaluation Boundary detector quality –Used for optimizing parameters –Comparing to other techniques Human-marked boundaries as ground truth –1000 images, 5-10 segmentations –Highly consistent

Evaluation Compare to ground truth using Precision-Recall –Sensitivity vs. Noise –Optimal tradeoff point is used for comparison – Fewer False Positives Fewer Misses Goal F 

Cue Calibration All free parameters optimized on training data Brightness Gradient –Disc Radius, bin/kernel sizes for KDE Color Gradient –Disc Radius, bin/kernel sizes for KDE, joint vs. marginals Texture Gradient –Disc Radius –Filter bank: scale, multiscale vs. singlescale –Histogram comparison method:  2, EMD, etc. –Number of textons (k-means value) –Image-specific vs. universal textons Localization parameters for each cue

Eliminate Redundant Cues Supervised learning using ground-truth images Orient energy (OE) same info as BG Multiple scales doesn’t add accuracy Best results –BG+TG+CG at single scale

Classifiers for Cue Combination Supervised learning using ground-truth images Logistic Regression –Linear and quadratic terms –Stable, quick, compact, intuitive –Training: MinutesEvaluation: Negligible compared to feature detection Density Estimation –Adaptive bins using k-means –Training: MinutesEvaluation: Negligible compared to feature detection Classification Trees –Top-down splits to maximize entropy, error bounded –Training: HoursEvaluation: Many x longer than regression Hierarchical Mixtures of Experts –8 experts, initialized top-down, fit with EM –Training: HoursEvaluation: 15 x longer than regression Support Vector Machines ( libsvm ) –Terrible: Model was large, exceedingly slow, brittle(parameters), opaque

Classifier Comparison

Cue Combinations

Two Decades of Boundary Detection

Pb Images I Canny2MMUsHumanImage

Pb Images II Canny2MMUsHumanImage

Pb Images III Canny2MMUsHumanImage

Summary & Comments Edge detection does not optimally segment natural images –Texture suppression is not sufficient Proposed method offers significant improvements –Simple but powerful feature detectors –Simple model for cue combination Empirical approach for calibration & cue combination Surprisingly effective for a low level approach Likely isn’t robust to larger textures –Offset by using multiple scales Prohibitively slow for on-line use –Minutes per image after optimizations Normalized Cuts: Boundaries  Regions