Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)

Slides:



Advertisements
Similar presentations
Primal-dual Algorithm for Convex Markov Random Fields Vladimir Kolmogorov University College London GDR (Optimisation Discrète, Graph Cuts et Analyse d'Images)
Advertisements

Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
Bayesian Belief Propagation
1 LP, extended maxflow, TRW OR: How to understand Vladimirs most recent work Ramin Zabih Cornell University.
Filling Algorithms Pixelwise MRFsChaos Mosaics Patch segments are pasted, overlapping, across the image. Then either: Ambiguities are removed by smoothing.
Various Regularization Methods in Computer Vision Min-Gyu Park Computer Vision Lab. School of Information and Communications GIST.
Graph Cut Algorithms for Computer Vision & Medical Imaging Ramin Zabih Computer Science & Radiology Cornell University Joint work with Y. Boykov, V. Kolmogorov,
Tutorial at ICCV (Barcelona, Spain, November 2011)
Introduction to Markov Random Fields and Graph Cuts Simon Prince
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
Computing Kemeny and Slater Rankings Vincent Conitzer (Joint work with Andrew Davenport and Jayant Kalagnanam at IBM Research.)
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
ICCV 2007 tutorial on Discrete Optimization Methods in Computer Vision part I Basic overview of graph cuts.
Optimal solution of binary problems Much material taken from :  Olga Veksler, University of Western Ontario
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
1 Fast Primal-Dual Strategies for MRF Optimization (Fast PD) Robot Perception Lab Taha Hamedani Aug 2014.
1 Minimum Ratio Contours For Meshes Andrew Clements Hao Zhang gruvi graphics + usability + visualization.
Corp. Research Princeton, NJ Computing geodesics and minimal surfaces via graph cuts Yuri Boykov, Siemens Research, Princeton, NJ joint work with Vladimir.
Optimal solution of binary problems Much material taken from :  Olga Veksler, University of Western Ontario
1 Can this be generalized?  NP-hard for Potts model [K/BVZ 01]  Two main approaches 1. Exact solution [Ishikawa 03] Large graph, convex V (arbitrary.
Semidefinite Programming
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
2010/5/171 Overview of graph cuts. 2010/5/172 Outline Introduction S-t Graph cuts Extension to multi-label problems Compare simulated annealing and alpha-
1 Computer Vision Research  Huttenlocher, Zabih –Recognition, stereopsis, restoration, learning  Strong algorithmic focus –Combinatorial optimization.
Stereo & Iterative Graph-Cuts Alex Rav-Acha Vision Course Hebrew University.
The max flow problem
Message Passing Algorithms for Optimization
Stereo Computation using Iterative Graph-Cuts
What Energy Functions Can be Minimized Using Graph Cuts? Shai Bagon Advanced Topics in Computer Vision June 2010.
Relaxations and Moves for MAP Estimation in MRFs M. Pawan Kumar STANFORDSTANFORD Vladimir KolmogorovPhilip TorrDaphne Koller.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
Graph-Cut Algorithm with Application to Computer Vision Presented by Yongsub Lim Applied Algorithm Laboratory.
Computer vision: models, learning and inference
Extensions of submodularity and their application in computer vision
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris École des Ponts ParisTech INRIA Saclay, Île-de-France Joint work with Phil.
Fast Approximate Energy Minimization via Graph Cuts
A Selective Overview of Graph Cut Energy Minimization Algorithms Ramin Zabih Computer Science Department Cornell University Joint work with Yuri Boykov,
Michael Bleyer LVA Stereo Vision
Nattee Niparnan. Easy & Hard Problem What is “difficulty” of problem? Difficult for computer scientist to derive algorithm for the problem? Difficult.
Graph Cut & Energy Minimization
Graph Cut Algorithms for Binocular Stereo with Occlusions
CS774. Markov Random Field : Theory and Application Lecture 13 Kyomin Jung KAIST Oct
Minimizing general submodular functions
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris Joint work with Phil Torr, Daphne Koller.
Lena Gorelick joint work with O. Veksler I. Ben Ayed A. Delong Y. Boykov.
Image segmentation Prof. Noah Snavely CS1114
Models in I.E. Lectures Introduction to Optimization Models: Shortest Paths.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
Fast and accurate energy minimization for static or time-varying Markov Random Fields (MRFs) Nikos Komodakis (Ecole Centrale Paris) Nikos Paragios (Ecole.
Lecture 19: Solving the Correspondence Problem with Graph Cuts CAP 5415 Fall 2006.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Tractable Higher Order Models in Computer Vision (Part II) Slides from Carsten Rother, Sebastian Nowozin, Pusohmeet Khli Microsoft Research Cambridge Presented.
Gaussian Mixture Models and Expectation-Maximization Algorithm.
Using Combinatorial Optimization within Max-Product Belief Propagation
A global approach Finding correspondence between a pair of epipolar lines for all pixels simultaneously Local method: no guarantee we will have one to.
Graph Algorithms for Vision Amy Gale November 5, 2002.
MAP Estimation in Binary MRFs using Bipartite Multi-Cuts Sashank J. Reddi Sunita Sarawagi Sundar Vishwanathan Indian Institute of Technology, Bombay TexPoint.
TU/e Algorithms (2IL15) – Lecture 12 1 Linear Programming.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
Markov Random Fields in Vision
The NP class. NP-completeness Lecture2. The NP-class The NP class is a class that contains all the problems that can be decided by a Non-Deterministic.
Energy functions f(p)  {0,1} : Ising model Can solve fast with graph cuts V( ,  ) = T[  ] : Potts model NP-hard Closely related to Multiway Cut Problem.
Rounding-based Moves for Metric Labeling M. Pawan Kumar École Centrale Paris INRIA Saclay, Île-de-France.
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
Markov Random Fields with Efficient Approximations
Efficient Graph Cut Optimization for Full CRFs with Quantized Edges
Discrete Optimization Methods Basic overview of graph cuts
Presentation transcript:

Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)

2 Schedule 9: :00: Overview (Zabih) 10: :10 Inference for learning (Zabih) 11: :30 More inference for learning, plus software demos (Komodakis, Kumar) 14: :00 Learning for inference (Komodakis) 16: :45 Advanced topics (Kumar) 17: :00 Discussion (all)

3 Recall: MAP-MRF energy function  Generalization of C2 is –Think of V as the cost for two adjacent pixels to have these particular labels –For binary images, the natural cost is uniform  Bayesian energy function: LikelihoodPrior

4 Generalizations  Many vision problems have this form  The optimization techniques are specific to these energy function, but not to images –See: [Kleinberg & Tardos JACM 02]

5 Network flow can help  For two labels, natural V is uniform –Ising model  Can write this as a minimizing a quadratic polynomial over binary variables  Problem can be solved exactly using network flow (thanks to Ford-Fulkerson) –Construction probably due to [Hammer et al. 65] –First applied to images by [Greig et al. 86]  Classical Computer Science problem reduction –Turn a new problem into a problem we can solve!

6 Important properties  Very efficient in practice –Lots of short paths, so roughly linear –Edmonds-Karp max flow algorithm finds augmenting paths in breadth-first order  Construction is symmetric (0 vs 1) –Each terminal is a label  Specific to 2 labels –Min cut with >2 terminals is NP-hard –Almost ANY such graph problem is NP-hard One has to appreciate Ford-Fulkerson!

7 Can this be generalized?  NP-hard for Potts model [K/BVZ 01]  Two main approaches 1. Exact solution [Ishikawa 03] Large graph, convex V (arbitrary D ) Not the considered the right prior for vision 2. Approximate solutions [BVZ 01] Solve a binary labeling problem, repeatedly Expansion move algorithm

8  Make green expansion move that most decreases E –Then make the best blue expansion move, etc –Done when no -expansion move decreases the energy, for any label  –See [BVZ 01] for details Input labeling f Green expansion move from f

9  Continuous vs. discrete –No floating point with graph cuts  Local min in line search vs. global min  Minimize over a line vs. hypersurface –Containing O(2 n ) candidates  Local minimum: weak vs. strong –Within 1% of global min on benchmarks! –Theoretical guarantees concerning distance from global minimum 2-approximation for a common choice of E Local improvement vs. Graph cuts

10 Binary sub-problem Input labeling Expansion moveBinary image

11 Expansion move energy Goal: find the binary image with lowest energy Binary image energy E ( b ) is restricted version of original E Depends on f, 

12 Regularity  The binary energy function is regular [KZ 04] if –Let’s all pretend this makes sense (for now)  Graph cuts solve regular functions!

13 When is binary energy regular?  Can find cheapest -expansion from f if  This is a Monge property –It holds if V is a metric, plus a few other cases  With this assumption, can compute the optimal expansion move

What does regularity mean?  Close relationship to graph problems solvable in polynomial time  Consider a set U and a real-valued function f defined on subsets of U –We say that f is submodular if: –Equivalent definition:  There are (slow) polynomial algorithms for submodular function minimization 14

Submodularity and graphs  Let U be the nodes of a weighted graph and f be the sum of outgoing edges –Called a cut function  Such a function is easily submodular –Same holds if the graph has terminals  Result: min cut is in P (barely)! 15

Submodularity and regularity  Consider the simplest non-trivial example –What does submodularity mean?  Now represent subsets as binary vectors –Regularity is simplest case of submodularity Not equivalent [Zivny], or we’d have a faster algorithm to minimize submodular functions 16

What is a regular QUBO?  Suppose we have a regular energy function. What does it look like as a polynomial?  Pick two pixels, write their variables  What is the cost involving these variables?  Regular if and only if quadratic term ≤ 0 17

18 Graph cuts and relaxations  Much current work uses relaxations to create better methods –Draws on work by P. Hammer and E. Boros, and their collaborators Quadratic Unconstrained Binary Optimization (QUBO) problems

19 The roof duality relaxation  Hammer, Boros et al solution –Graph construction called “roof duality” due to [Hammer et al 84] [Boros & Hammer 02] –Introduced into computer vision by V. Kolmogorov in early 2005 Efficient implementation often called QPBO  Relaxation with a very nice property –Lots of useful information even if it doesn’t compute the optimal expansion move –Partial optimality! Right answers at some pixels

Opt (y=2) Partial optimality is hard! False Opt (y=1)

21 Roof duality relaxation  Alternate encoding of expansion moves  Can’t use graph cuts to minimize  But we can minimize the relaxation –Note:

22 Partial optimality  [Hammer et al 84][Boros & Hammer 02] show this relaxation has an amazing property –Strong persistency: all consistent pixels have the correct label –If there are no inconsistent pixels (or very few), this is great Which happens often in MR! ?

23 Recent MRI results SENSE (= LS) Graph cuts

24 How to get partial optimality?  We’re trying to minimize an arbitrary binary function (NP-hard) –How to find some values at global min?  Suppose we could always decrease the energy by setting (binary) variable a=1 –Then we know a=1 in global min Or we could go there, then decrease energy! –Use max flow to find such a partial assignment Typically on multiple variables

Posiforms  Can rewrite our function so that all (non- constant) coefficients are positive –Always pay K. Pay 2 if a 1 =1. Pay 4 if a 2 =0…  How can we find a partial assignment that always reduces the cost? –Generalization of “pure literal rule”  Max flow finds such partial assignments! 25

Probing  Invented by Boros et al, introduced into vision by Rother et al CVPR07  Another use of “consensus”, aka resolution (A  B)  (-A  B)  B 26

Higher order priors?  Problem: We can solve quadratic binary energy functions, but we need to reduce non-quadratic priors to quadratic ones  Ishikawa [TPAMI10] gave the first practical algorithm to do this in all cases –Previously could only handle negative terms (Freedman & Drineas, generalizing [Kolmogorov & Zabih TPAMI04]) 27

Recent progress  “A Graph Cut Algorithm for Higher-order Markov Random Fields”, A. Fix, A. Gruber, E. Boros, & R. Zabih, ICCV11 poster  Significantly improve Ishikawa’s results  Characterizes a class of problems where our performance is even better –shows that some common computer vision problems are in this class 28

ICCV11 poster preview  Positive terms are the hard case  Ishikawa does them one at a time, we do a different reduction for groups sharing a common variable –Works especially well on vision problems 29

Results 30

31 Conclusions  Flow-based algorithms (“graph cuts”) are powerful tools for MAP estimation of MRF’s –Common in computer vision, and elsewhere  Lots of interest in using these methods for even broader classes of problems  Graph cuts can give strong results for linear inverse systems with discontinuities –There are many of these (in MR and beyond)