Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)

Slides:



Advertisements
Similar presentations
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London Tutorial at GDR (Optimisation Discrète, Graph Cuts.
Advertisements

Bayesian Belief Propagation
1 LP, extended maxflow, TRW OR: How to understand Vladimirs most recent work Ramin Zabih Cornell University.
Graph Cut Algorithms for Computer Vision & Medical Imaging Ramin Zabih Computer Science & Radiology Cornell University Joint work with Y. Boykov, V. Kolmogorov,
Least squares CS1114
Learning to Combine Bottom-Up and Top-Down Segmentation Anat Levin and Yair Weiss School of CS&Eng, The Hebrew University of Jerusalem, Israel.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
I Images as graphs Fully-connected graph – node for every pixel – link between every pair of pixels, p,q – similarity w ij for each link j w ij c Source:
The University of Ontario CS 4487/9687 Algorithms for Image Analysis Multi-Label Image Analysis Problems.
Proportion Priors for Image Sequence Segmentation Claudia Nieuwenhuis, etc. ICCV 2013 Oral.
Variational Inference in Bayesian Submodular Models
Discrete Optimization for Vision and Learning. Who? How? M. Pawan Kumar Associate Professor Ecole Centrale Paris Nikos Komodakis Associate Professor Ecole.
Optimal solution of binary problems Much material taken from :  Olga Veksler, University of Western Ontario
Graph-Based Image Segmentation
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)
1 Fast Primal-Dual Strategies for MRF Optimization (Fast PD) Robot Perception Lab Taha Hamedani Aug 2014.
Markov Random Fields (MRF)
Prof. Ramin Zabih (CS) Prof. Ashish Raj (Radiology) CS5540: Computational Techniques for Analyzing Clinical Data.
Optimal solution of binary problems Much material taken from :  Olga Veksler, University of Western Ontario
1 Can this be generalized?  NP-hard for Potts model [K/BVZ 01]  Two main approaches 1. Exact solution [Ishikawa 03] Large graph, convex V (arbitrary.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
2010/5/171 Overview of graph cuts. 2010/5/172 Outline Introduction S-t Graph cuts Extension to multi-label problems Compare simulated annealing and alpha-
A Closed Form Solution to Natural Image Matting
Stereo & Iterative Graph-Cuts Alex Rav-Acha Vision Course Hebrew University.
The plan for today Camera matrix
Announcements Readings for today:
Lecture 10: Stereo and Graph Cuts
Stereo Computation using Iterative Graph-Cuts
An Iterative Optimization Approach for Unified Image Segmentation and Matting Hello everyone, my name is Jue Wang, I’m glad to be here to present our paper.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
Announcements PS3 Due Thursday PS4 Available today, due 4/17. Quiz 2 4/24.
Stereo Matching & Energy Minimization Vision for Graphics CSE 590SS, Winter 2001 Richard Szeliski.
Computer Vision Optical Flow Marc Pollefeys COMP 256 Some slides and illustrations from L. Van Gool, T. Darell, B. Horn, Y. Weiss, P. Anandan, M. Black,
Crash Course on Machine Learning
Introduction to Monte Carlo Methods D.J.C. Mackay.
Efficient Algorithms for Matching Pedro Felzenszwalb Trevor Darrell Yann LeCun Alex Berg.
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/31/15.
A Selective Overview of Graph Cut Energy Minimization Algorithms Ramin Zabih Computer Science Department Cornell University Joint work with Yuri Boykov,
Graph-based Segmentation. Main Ideas Convert image into a graph Vertices for the pixels Vertices for the pixels Edges between the pixels Edges between.
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/24/10.
Machine Learning Chapter 4. Artificial Neural Networks
City University of Hong Kong 18 th Intl. Conf. Pattern Recognition Self-Validated and Spatially Coherent Clustering with NS-MRF and Graph Cuts Wei Feng.
Planar Cycle Covering Graphs for inference in MRFS The Typhon Algorithm A New Variational Approach to Ground State Computation in Binary Planar Markov.
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris Joint work with Phil Torr, Daphne Koller.
Graph Cuts Marc Niethammer. Segmentation by Graph-Cuts A way to compute solutions to the optimization problems we looked at before. Example: Binary Segmentation.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Discrete Optimization Lecture 3 – Part 1 M. Pawan Kumar Slides available online
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
Probabilistic Inference Lecture 5 M. Pawan Kumar Slides available online
Dynamic Tree Block Coordinate Ascent Daniel Tarlow 1, Dhruv Batra 2 Pushmeet Kohli 3, Vladimir Kolmogorov 4 1: University of Toronto3: Microsoft Research.
Lecture 19: Solving the Correspondence Problem with Graph Cuts CAP 5415 Fall 2006.
Clustering Prof. Ramin Zabih
A global approach Finding correspondence between a pair of epipolar lines for all pixels simultaneously Local method: no guarantee we will have one to.
Graph Algorithms for Vision Amy Gale November 5, 2002.
Photoconsistency constraint C2 q C1 p l = 2 l = 3 Depth labels If this 3D point is visible in both cameras, pixels p and q should have similar intensities.
Markov Random Fields in Vision
Energy minimization Another global approach to improve quality of correspondences Assumption: disparities vary (mostly) smoothly Minimize energy function:
MRFs and Segmentation with Graph Cuts Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/27/12.
Jungwook Choi and Rob A. Rutenbar
Alexander Shekhovtsov and Václav Hlaváč
Heuristic Optimization Methods
Markov Random Fields with Efficient Approximations
Efficient Graph Cut Optimization for Full CRFs with Quantized Edges
Learning to Combine Bottom-Up and Top-Down Segmentation
Lecture 31: Graph-Based Image Segmentation
Announcements more panorama slots available now
Expectation-Maximization & Belief Propagation
Announcements more panorama slots available now
“Traditional” image segmentation
Presentation transcript:

Learning with Inference for Discrete Graphical Models Nikos Komodakis Pawan Kumar Nikos Paragios Ramin Zabih (presenter)

2 Schedule 9: :00: Overview (Zabih) 10: :10 Inference for learning (Zabih) 11: :30 More inference for learning, plus software demos (Komodakis, Kumar) 14: :00 Learning for inference (Komodakis) 16: :45 Advanced topics (Kumar) 17: :00 Discussion (all)

3 Overview

4 Motivating example  Suppose we want to find a bright object against a dark background –But some of the pixel values are slightly wrong

5 Optimization viewpoint  Find best (least expensive) binary image –Costs: C1 (labeling) and C2 (boundary)  C1: Labeling a dark pixel as foreground –Or, a bright pixel as background  If we only had labeling costs, the cheapest solution is the thresholded output  C2: The length of the boundary between foreground and background –Penalizes isolated pixels or ragged boundaries

6 MAP-MRF energy function  Generalization of C2 is –Think of V as the cost for two adjacent pixels to have these particular labels –For binary images, the natural cost is uniform  Bayesian energy function: LikelihoodPrior

Historical view  Energy functions like this go back at least as far as Horn & Schunk (1981)  The Bayesian view was popularized by Geman and Geman (TPAMI 1984)  Historically solved by gradient descent or related methods (e.g. annealing) –Optimization method and energy function are not independent choices! –Use the most specific method you can And, be prepared to tweak your problem 7

8 Discrete methods  Starting in the late 90’s researchers (re-) discovered discrete optimization methods –Graph cuts, belief prop, dynamic programming, linear programming, semi- definite programming, etc.  These methods proved remarkably effective at solving problems that could not be solved before  Vision has lots of cool math –interest in this area is largely driven by performance!

Performance overview  Best summary: Szeliski et al. “A comparative study of energy minimization methods for Markov Random Fields with smoothness-based priors”, TPAMI 2008 –An updated version is a chapter in “Markov Random Fields for Vision and Image Processing”, 2011  LP-based methods compute lower bounds –Use this to measure performance 9

Typical results 10

11 Correlation Graph cuts Stereo images Right answers

Is vision solved? Can we all go home now?  For many easy problems the technical problem of minimizing the energy is now effectively solved –“Easy” = “submodular/regular, & first-order” We’ll define these terms later on –“Technical problem” ≠ vision problem –“The energy”? Is the right one obvious??  Still, this is vast progress in a relatively short period of time –These “easy” problems were impossible in ‘97! 12

What is the right energy?  Sometimes we can find the global optimum fast –Original example can be solved by graph cuts  Do we get what we want? –How important is C1 (data) vs C2 (prior)? –If C2 dominates, we get a uniform image  Important lessons –Need to learn the right parameter values –Prior is not actually strong enough 13

Better priors?  Original graph cuts example, from Greig et al 1989 (example from Olga Veksler)  No choice of the relative importance of C1 and C2 gives the letter A at global min! 14

How good is global min?  We can often get a solution whose energy is lower than the ground truth –Folk theorem, first published in [Tappen & Freedman ICCV03], improved by [Meltzer, Yanover & Weiss ICCV05] –Huge gap! Can easily be 40% or more  Lots of parameters in energy functions –Need to learn them –Pretty clear that priors with fast algorithms are just too weak for our purposes 15

Learning and inference  How does learning come into play? –There are too many parameters to an energy function to tune by hand Example: Felzenszwalb deformable parts-based models have thousands of parameters  Two topics for this afternoon –Parameter estimation can be formulated as an optimization problem –We need methods that can learn parameters from real data, with all its imperfections 16