Learning to Combine Bottom-Up and Top-Down Segmentation

Slides:



Advertisements
Similar presentations
POSE–CUT Simultaneous Segmentation and 3D Pose Estimation of Humans using Dynamic Graph Cuts Mathieu Bray Pushmeet Kohli Philip H.S. Torr Department of.
Advertisements

Bayesian Belief Propagation
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
HOPS: Efficient Region Labeling using Higher Order Proxy Neighborhoods Albert Y. C. Chen 1, Jason J. Corso 1, and Le Wang 2 1 Dept. of Computer Science.
The Layout Consistent Random Field for detecting and segmenting occluded objects CVPR, June 2006 John Winn Jamie Shotton.
Top-Down & Bottom-Up Segmentation
Constrained Approximate Maximum Entropy Learning (CAMEL) Varun Ganapathi, David Vickrey, John Duchi, Daphne Koller Stanford University TexPoint fonts used.
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Part 4: Combined segmentation and recognition by Rob Fergus (MIT)
Learning to Combine Bottom-Up and Top-Down Segmentation Anat Levin and Yair Weiss School of CS&Eng, The Hebrew University of Jerusalem, Israel.
Scene Labeling Using Beam Search Under Mutex Constraints ID: O-2B-6 Anirban Roy and Sinisa Todorovic Oregon State University 1.
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
1School of CS&Eng The Hebrew University
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
LOCUS (Learning Object Classes with Unsupervised Segmentation) A variational approach to learning model- based segmentation. John Winn Microsoft Research.
Contour Based Approaches for Visual Object Recognition Jamie Shotton University of Cambridge Joint work with Roberto Cipolla, Andrew Blake.
Biased Normalized Cuts 1 Subhransu Maji and Jithndra Malik University of California, Berkeley IEEE Conference on Computer Vision and Pattern Recognition.
Visual Recognition Tutorial
A Graphical Model For Simultaneous Partitioning And Labeling Philip Cowans & Martin Szummer AISTATS, Jan 2005 Cambridge.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Robust Higher Order Potentials For Enforcing Label Consistency
Schedule Introduction Models: small cliques and special potentials Tea break Inference: Relaxation techniques:
LARGE-SCALE NONPARAMETRIC IMAGE PARSING Joseph Tighe and Svetlana Lazebnik University of North Carolina at Chapel Hill CVPR 2011Workshop on Large-Scale.
Abstract We present a model of curvilinear grouping using piecewise linear representations of contours and a conditional random field to capture continuity.
A Closed Form Solution to Natural Image Matting
TextonBoost : Joint Appearance, Shape and Context Modeling for Multi-Class Object Recognition and Segmentation J. Shotton*, J. Winn†, C. Rother†, and A.
Oxford Brookes Seminar Thursday 3 rd September, 2009 University College London1 Representing Object-level Knowledge for Segmentation and Image Parsing:
CVR05 University of California Berkeley 1 Cue Integration in Figure/Ground Labeling Xiaofeng Ren, Charless Fowlkes, Jitendra Malik.
WORD-PREDICTION AS A TOOL TO EVALUATE LOW-LEVEL VISION PROCESSES Prasad Gabbur, Kobus Barnard University of Arizona.
1 Segmentation with Scene and Sub-Scene Categories Joseph Djugash Input Image Scene/Sub-Scene Classification Segmentation.
The Layout Consistent Random Field for Recognizing and Segmenting Partially Occluded Objects By John Winn & Jamie Shotton CVPR 2006 presented by Tomasz.
Computer Vision Group University of California Berkeley 1 Cue Integration in Figure/Ground Labeling Xiaofeng Ren, Charless Fowlkes and Jitendra Malik.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
Cue Integration in Figure/Ground Labeling Xiaofeng Ren, Charless Fowlkes and Jitendra Malik, U.C. Berkeley We present a model of edge and region grouping.
What, Where & How Many? Combining Object Detectors and CRFs
A Trainable Graph Combination Scheme for Belief Propagation Kai Ju Liu New York University.
Crash Course on Machine Learning
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
1 Naïve Bayes Models for Probability Estimation Daniel Lowd University of Washington (Joint work with Pedro Domingos)
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/24/10.
1 Contours and Junctions in Natural Images Jitendra Malik University of California at Berkeley (with Jianbo Shi, Thomas Leung, Serge Belongie, Charless.
Segmentation slides adopted from Svetlana Lazebnik.
Recognition using Boosting Modified from various sources including
Leo Zhu CSAIL MIT Joint work with Chen, Yuille, Freeman and Torralba 1.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Texture We would like to thank Amnon Drory for this deck הבהרה : החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall 2004 AdaBoost.. Binary Classification. Read 9.5 Duda,
Associative Hierarchical CRFs for Object Class Image Segmentation
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
Object Recognition by Integrating Multiple Image Segmentations Caroline Pantofaru, Cordelia Schmid, Martial Hebert ECCV 2008 E.
TextonBoost: Joint Appearance, Shape and Context Modeling for Multi-Class Object Recognition and Segmentation J. Shotton ; University of Cambridge J. Jinn,
Part 4: combined segmentation and recognition Li Fei-Fei.
Image segmentation.
Segmentation slides adopted from Svetlana Lazebnik.
Holistic Scene Understanding Virginia Tech ECE /02/26 Stanislaw Antol.
MRFs (X1,X2) X3 X1 X2 4 (X2,X3,X3) X4. MRFs (X1,X2) X3 X1 X2 4 (X2,X3,X3) X4.
Learning Deep Generative Models by Ruslan Salakhutdinov
Learning a Region-based Scene Segmentation Model
Classification of unlabeled data:
LOCUS: Learning Object Classes with Unsupervised Segmentation
Nonparametric Semantic Segmentation
Saliency detection Donghun Yeo CV Lab..
Integration and Graphical Models
المشرف د.يــــاســـــــــر فـــــــؤاد By: ahmed badrealldeen
Markov Random Fields Presented by: Vladan Radosavljevic.
Expectation-Maximization & Belief Propagation
The EM Algorithm With Applications To Image Epitome
“Traditional” image segmentation
Presentation transcript:

Learning to Combine Bottom-Up and Top-Down Segmentation Anat Levin and Yair Weiss School of CS&Eng, The Hebrew University of Jerusalem, Israel

Bottom-up segmentation Bottom-up approaches: Use low level cues to group similar pixels Malik et al, 2000 Sharon et al, 2001 Comaniciu and Meer, 2002 …

Bottom-up segmentation is ill posed Many possible segmentation are equally good based on low level cues alone. Some segmentation example (maybe horses from Eran’s paper) images from Borenstein and Ullman 02

Top-down segmentation Class-specific, top-down segmentation (Borenstein & Ullman Eccv02) Winn and Jojic 05 Leibe et al 04 Yuille and Hallinan 02. Liu and Sclaroff 01 Yu and Shi 03

Combining top-down and bottom-up segmentation + Find a segmentation: Similar to the top-down model Aligns with image edges

Previous approaches Borenstein et al 04 Combining top-down and bottom up segmentation. Tu et al ICCV03 Image parsing: segmentation, detection, and recognition. Kumar et al CVPR05 Obj-Cut. Shotton et al ECCV06: TextonBoost Previous approaches: Train top-down and bottom-up models independently

Why learning top-down and bottom-up models simultaneously? Large number of freedom degrees in tentacles configuration- requires a complex deformable top down model On the other hand: rather uniform colors- low level segmentation is easy

Our approach Learn top-down and bottom-up models simultaneously Reduces at run time to energy minimization with binary labels (graph min cut)

Segmentation alignment with image edges Energy model Segmentation alignment with image edges Consistency with fragments segmentation

Segmentation alignment with image edges Energy model Segmentation alignment with image edges Consistency with fragments segmentation

Segmentation alignment with image edges Energy model Segmentation alignment with image edges Consistency with fragments segmentation Resulting min-cut segmentation

Learning from segmented class images Training data: Goal: Learn fragments for an energy function

Learning energy functions using conditional random fields Theory of CRFs: Lafferty et al 2001 LeCun and Huang 2005 CRFs For vision: Kumar and Hebert 2003 Ren et al 2006 He et al 2004, 2006 Quattoni et al 2005 Torralba et al 04

Learning energy functions using conditional random fields Maximize energy of all other configurations Minimize energy of true segmentation E(x) “It's not enough to succeed. Others must fail.” –Gore Vidal

Learning energy functions using conditional random fields Maximize energy of all other configurations Minimize energy of true segmentation P(x) P(x) “It's not enough to succeed. Others must fail.” –Gore Vidal

Differentiating CRFs log-likelihood Log-likelihood is convex with respect to Log-likelihood gradients with respect to : Expected feature response minus observed feature response Yair- in the original version of this slide I had another equation expressing the expectation as a sum of marginals (see next hidden slide). At least for me, it wasn’t originally clear what this expectation means before I saw the other equation. However, I try to delete un necessary equations..

Conditional random fields-computational challenges CRFs cost- evaluating partition function Derivatives- evaluating marginal probabilities Use approximate estimations: Sampling Belief Propagation and Bethe free energy Used in this work: Tree reweighted belief propagation and Tree reweighted upper bound (Wainwright et al 03)

Fragments selection Candidate fragments pool: Greedy energy design:

Fragments selection challenges Straightforward computation of likelihood improvement is impractical 2000 Fragments 50 Training images 10 Fragments selection iterations 1,000,000 inference operations!

Fragment with low error on the training set Fragments selection First order approximation to log-likelihood gain: Fragment with low error on the training set Fragment not accounted for by the existing model Similar idea in different contexts: Zhu et al 1997 Lafferty et al 2004 McCallum 2003

Fragments selection First order approximation to log-likelihood gain: Requires a single inference process on the previous iteration energy to evaluate approximations with respect to all fragments First order approximation evaluation is linear in the fragment size

Fragments selection- summary Initialization: Low- level term For k=1:K Run TRBP inference using the previous iteration energy. Approximate likelihood gain of candidate fragments Add to energy the fragment with maximal gain.

Training horses model

Training horses model-one fragment

Training horses model-two fragments

Training horses model-three fragments

Results- horses dataset

Results- horses dataset Mislabeled pixels percent Fragments number Comparable to previous results (Kumar et al, Borenstein et al.) but with far fewer fragments

Results- artificial octopi

From the TU Darmstadt Database Results- cows dataset From the TU Darmstadt Database

Results- cows dataset Mislabeled pixels percent Fragments number

Conclusions Simultaneously learning top-down and bottom-up segmentation cues. Learning formulated as estimation in Conditional Random Fields Novel, efficient fragments selection algorithm Algorithm achieves state of the art performance with a significantly smaller number of fragments