An Analysis of Convex Relaxations (PART I) Minimizing Higher Order Energy Functions (PART 2) Philip Torr Work in collaboration with: Pushmeet Kohli, Srikumar.

Slides:



Advertisements
Similar presentations
POSE–CUT Simultaneous Segmentation and 3D Pose Estimation of Humans using Dynamic Graph Cuts Mathieu Bray Pushmeet Kohli Philip H.S. Torr Department of.
Advertisements

Mean-Field Theory and Its Applications In Computer Vision1 1.
1 LP, extended maxflow, TRW OR: How to understand Vladimirs most recent work Ramin Zabih Cornell University.
O BJ C UT M. Pawan Kumar Philip Torr Andrew Zisserman UNIVERSITY OF OXFORD.
Solving Markov Random Fields using Second Order Cone Programming Relaxations M. Pawan Kumar Philip Torr Andrew Zisserman.
Solving Markov Random Fields using Dynamic Graph Cuts & Second Order Cone Programming Relaxations M. Pawan Kumar, Pushmeet Kohli Philip Torr.
Convex Programming Brookes Vision Reading Group. Huh? What is convex ??? What is programming ??? What is convex programming ???
Engineering Optimization
1 OR II GSLM Outline  some terminology  differences between LP and NLP  basic questions in NLP  gradient and Hessian  quadratic form  contour,
Introduction to Markov Random Fields and Graph Cuts Simon Prince
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
by Rianto Adhy Sasongko Supervisor: Dr.J.C.Allwright
Continuous optimization Problems and successes
Efficient Inference for Fully-Connected CRFs with Stationarity
1 Fast Primal-Dual Strategies for MRF Optimization (Fast PD) Robot Perception Lab Taha Hamedani Aug 2014.
Approximation Algoirthms: Semidefinite Programming Lecture 19: Mar 22.
Graph Algorithms: Minimum Spanning Tree We are given a weighted, undirected graph G = (V, E), with weight function w:
Semidefinite Programming
ICCV Tutorial 2007 Philip Torr Papers, presentations and videos on web.....
An Analysis of Convex Relaxations M. Pawan Kumar Vladimir Kolmogorov Philip Torr for MAP Estimation.
P 3 & Beyond Solving Energies with Higher Order Cliques Pushmeet Kohli Pawan Kumar Philip H. S. Torr Oxford Brookes University CVPR 2007.
Improved Moves for Truncated Convex Models M. Pawan Kumar Philip Torr.
Efficiently Solving Convex Relaxations M. Pawan Kumar University of Oxford for MAP Estimation Philip Torr Oxford Brookes University.
Exploiting Duality (Particularly the dual of SVM) M. Pawan Kumar VISUAL GEOMETRY GROUP.
Graph Cut based Inference with Co-occurrence Statistics Ľubor Ladický, Chris Russell, Pushmeet Kohli, Philip Torr.
Relaxations and Moves for MAP Estimation in MRFs M. Pawan Kumar STANFORDSTANFORD Vladimir KolmogorovPhilip TorrDaphne Koller.
Protein Encoding Optimization Student: Logan Everett Mentor: Endre Boros Funded by DIMACS REU 2004.
Hierarchical Graph Cuts for Semi-Metric Labeling M. Pawan Kumar Joint work with Daphne Koller.
Invariant Large Margin Nearest Neighbour Classifier M. Pawan Kumar Philip Torr Andrew Zisserman.
Measuring Uncertainty in Graph Cut Solutions Pushmeet Kohli Philip H.S. Torr Department of Computing Oxford Brookes University.
ECE Synthesis & Verification - LP Scheduling 1 ECE 667 ECE 667 Synthesis and Verification of Digital Circuits Scheduling Algorithms Analytical approach.
MAP Estimation Algorithms in M. Pawan Kumar, University of Oxford Pushmeet Kohli, Microsoft Research Computer Vision - Part I.
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris École des Ponts ParisTech INRIA Saclay, Île-de-France Joint work with Phil.
An Invariant Large Margin Nearest Neighbour Classifier Results Matching Faces from TV Video Aim: To learn a distance metric for invariant nearest neighbour.
Probabilistic Inference Lecture 4 – Part 2 M. Pawan Kumar Slides available online
CS774. Markov Random Field : Theory and Application Lecture 08 Kyomin Jung KAIST Sep
Multiplicative Bounds for Metric Labeling M. Pawan Kumar École Centrale Paris Joint work with Phil Torr, Daphne Koller.
Rounding-based Moves for Metric Labeling M. Pawan Kumar Center for Visual Computing Ecole Centrale Paris.
Learning a Small Mixture of Trees M. Pawan Kumar Daphne Koller Aim: To efficiently learn a.
Discrete Optimization Lecture 2 – Part I M. Pawan Kumar Slides available online
Probabilistic Inference Lecture 3 M. Pawan Kumar Slides available online
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Discrete Optimization in Computer Vision M. Pawan Kumar Slides will be available online
Nonlinear Programming Models
Discrete Optimization Lecture 3 – Part 1 M. Pawan Kumar Slides available online
Probabilistic Inference Lecture 5 M. Pawan Kumar Slides available online
Efficient Discriminative Learning of Parts-based Models M. Pawan Kumar Andrew Zisserman Philip Torr
Associative Hierarchical CRFs for Object Class Image Segmentation
Introduction to Semidefinite Programs Masakazu Kojima Semidefinite Programming and Its Applications Institute for Mathematical Sciences National University.
Robust Optimization and Applications Laurent El Ghaoui IMA Tutorial, March 11, 2003.
Tractable Higher Order Models in Computer Vision (Part II) Slides from Carsten Rother, Sebastian Nowozin, Pusohmeet Khli Microsoft Research Cambridge Presented.
Discrete Optimization Lecture 2 – Part 2 M. Pawan Kumar Slides available online
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
Inference for Learning Belief Propagation. So far... Exact methods for submodular energies Approximations for non-submodular energies Move-making ( N_Variables.
Probabilistic Inference Lecture 2 M. Pawan Kumar Slides available online
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Unique Games Approximation Amit Weinstein Complexity Seminar, Fall 2006 Based on: “Near Optimal Algorithms for Unique Games" by M. Charikar, K. Makarychev,
Discrete Optimization Lecture 1 M. Pawan Kumar Slides available online
Tightening LP Relaxations for MAP using Message-Passing David Sontag Joint work with Talya Meltzer, Amir Globerson, Tommi Jaakkola, and Yair Weiss.
MAP Estimation in Binary MRFs using Bipartite Multi-Cuts Sashank J. Reddi Sunita Sarawagi Sundar Vishwanathan Indian Institute of Technology, Bombay TexPoint.
1 Ratio-Based Efficiency Analysis (REA) Antti Punkka and Ahti Salo Systems Analysis Laboratory Aalto University School of Science and Technology P.O. Box.
MAP Estimation of Semi-Metric MRFs via Hierarchical Graph Cuts M. Pawan Kumar Daphne Koller Aim: To obtain accurate, efficient maximum a posteriori (MAP)
Optimization - Lecture 5, Part 1 M. Pawan Kumar Slides available online
TU/e Algorithms (2IL15) – Lecture 12 1 Linear Programming.
Linear Programming Chap 2. The Geometry of LP  In the text, polyhedron is defined as P = { x  R n : Ax  b }. So some of our earlier results should.
Rounding-based Moves for Metric Labeling M. Pawan Kumar École Centrale Paris INRIA Saclay, Île-de-France.
Lecture 8 – Nonlinear Programming Models
An Analysis of Convex Relaxations for MAP Estimation
Presentation transcript:

An Analysis of Convex Relaxations (PART I) Minimizing Higher Order Energy Functions (PART 2) Philip Torr Work in collaboration with: Pushmeet Kohli, Srikumar Ramalingam M. Pawan Kumar, Ľubor Ladický, Vladimir Kolmogorov

An Analysis of Convex Relaxations M. Pawan Kumar Vladimir Kolmogorov Philip Torr for MAP Estimation

Aim V1V1 V2V2 V3V3 V4V4 Label ‘0’ Label ‘1’ Labelling m = {1, 0, 0, 1} Random Variables V = {V 1,...,V 4 } Label Set L = {0, 1} To analyze convex relaxations for MAP estimation

Aim V1V1 V2V2 V3V3 V4V4 Label ‘0’ Label ‘1’ Cost(m) = = 13 Which approximate algorithm is the best? Minimum Cost Labelling? NP-hard problem To analyze convex relaxations for MAP estimation

Aim V1V1 V2V2 V3V3 V4V4 Label ‘0’ Label ‘1’ Objectives Compare existing convex relaxations – LP, QP and SOCP Develop new relaxations based on the comparison To analyze convex relaxations for MAP estimation

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results Two New SOCP Relaxations

Integer Programming Formulation V1V1 V2V2 Label ‘0’ Label ‘1’ Unary Cost Unary Cost Vector u = [ 5 Cost of V 1 = 0 2 Cost of V 1 = 1 ; 2 4 ] Labelling m = {1, 0}

V1V1 V2V2 Label ‘0’ Label ‘1’ Unary Cost Unary Cost Vector u = [ 5 2 ; 2 4 ] T Labelling m = {1, 0} Label vector x = [ -1 V 1  0 1 V 1 = 1 ; 1 -1 ] T Recall that the aim is to find the optimal x Integer Programming Formulation

V1V1 V2V2 Label ‘0’ Label ‘1’ Unary Cost Unary Cost Vector u = [ 5 2 ; 2 4 ] T Labelling m = {1, 0} Label vector x = [ -11; 1 -1 ] T Sum of Unary Costs = 1 2 ∑ i u i (1 + x i ) Integer Programming Formulation

V1V1 V2V2 Label ‘0’ Label ‘1’ Pairwise Cost Labelling m = {1, 0} 0 Cost of V 1 = 0 and V 1 = Cost of V 1 = 0 and V 2 = 0 3 Cost of V 1 = 0 and V 2 = Pairwise Cost Matrix P Integer Programming Formulation

V1V1 V2V2 Label ‘0’ Label ‘1’ Pairwise Cost Labelling m = {1, 0} Pairwise Cost Matrix P Sum of Pairwise Costs 1 4 ∑ ij P ij (1 + x i )(1+x j ) Integer Programming Formulation

V1V1 V2V2 Label ‘0’ Label ‘1’ Pairwise Cost Labelling m = {1, 0} Pairwise Cost Matrix P Sum of Pairwise Costs 1 4 ∑ ij P ij (1 + x i +x j + x i x j ) 1 4 ∑ ij P ij (1 + x i + x j + X ij )= X = x x T X ij = x i x j Integer Programming Formulation

Constraints Uniqueness Constraint ∑ x i = 2 - |L| i  V a Integer Constraints x i  {-1,1} X = x x T Integer Programming Formulation

x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  {-1,1} X = x x T Convex Non-Convex Integer Programming Formulation

Outline Integer Programming Formulation Existing Relaxations –Linear Programming (LP-S) –Semidefinite Programming (SDP-L) –Second Order Cone Programming (SOCP-MS) Comparison Generalization of Results Two New SOCP Relaxations

LP-S x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  {-1,1} X = x x T Retain Convex Part Schlesinger, 1976 Relax Non-Convex Constraint

LP-S x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] X = x x T Retain Convex Part Schlesinger, 1976 Relax Non-Convex Constraint

LP-S x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a Retain Convex Part Schlesinger, 1976 X ij  [-1,1] 1 + x i + x j + X ij ≥ 0 ∑ X ij = (2 - |L|) x i j  V b LP-S x i  [-1,1]

Outline Integer Programming Formulation Existing Relaxations –Linear Programming (LP-S) –Semidefinite Programming (SDP-L) –Second Order Cone Programming (SOCP-MS) Comparison Generalization of Results Two New SOCP Relaxations Experiments

SDP-L x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  {-1,1} X = x x T Retain Convex Part Lasserre, 2000 Relax Non-Convex Constraint

SDP-L x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] X = x x T Retain Convex Part Relax Non-Convex Constraint Lasserre, 2000

SDP-L x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] Retain Convex Part X ii = 1 X - xx T 0 Accurate Inefficient Lasserre, 2000

Outline Integer Programming Formulation Existing Relaxations –Linear Programming (LP-S) –Semidefinite Programming (SDP-L) –Second Order Cone Programming (SOCP-MS) Comparison Generalization of Results Two New SOCP Relaxations

SOCP Relaxation x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] X ii = 1 X - xx T 0 Derive SOCP relaxation from the SDP relaxation Further Relaxation

2-D Example X 11 X 12 X 21 X 22 1X 12 1 = X = x1x1x1x1 x1x2x1x2 x2x1x2x1 x2x2x2x2 xx T = x12x12 x1x2x1x2 x1x2x1x2 = x22x22

2-D Example (X - xx T ) 1 - x 1 2 X 12 -x 1 x x 2 2 C 1  0 C 1 0 (x 1 + x 2 ) 2  2 + 2X 12 SOC of the form || v || 2  st 

2-D Example (X - xx T ) 1 - x 1 2 X 12 -x 1 x x 2 2 C 2  0 C 2 0 (x 1 - x 2 ) 2  2 - 2X 12 SOC of the form || v || 2  st  0 1 1

SOCP Relaxation Consider a matrix C 1 = UU T 0 (X - xx T ) ||U T x || 2  X. C 1 C 1.  0 Continue for C 2, C 3, …, C n SOC of the form || v || 2  st Kim and Kojima, 2000

SOCP Relaxation How many constraints for SOCP = SDP ? Infinite. Specify constraints similar to the 2-D example xixi xjxj X ij (x i + x j ) 2  2 + 2X ij (x i + x j ) 2  2 - 2X ij

SOCP-MS x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] X ii = 1 X - xx T 0 Muramatsu and Suzuki, 2003

SOCP-MS x* = argmin 1 2 ∑ u i (1 + x i ) ∑ P ij (1 + x i + x j + X ij ) ∑ x i = 2 - |L| i  V a x i  [-1,1] (x i + x j ) 2  2 + 2X ij (x i - x j ) 2  2 - 2X ij Specified only when P ij  0 Muramatsu and Suzuki, 2003

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results Two New SOCP Relaxations

Dominating Relaxation For all MAP Estimation problem (u, P) A dominates B A B ≥ Dominating relaxations are better

Equivalent Relaxations A dominates B B dominates A Strictly Dominating Relaxation A dominates B B does not dominate A

SOCP-MS (x i + x j ) 2  2 + 2X ij (x i - x j ) 2  2 - 2X ij min  ij P ij X ij P ij ≥ 0 (x i + x j ) X ij = P ij < 0 (x i - x j ) X ij = SOCP-MS is a QP Same as QP by Ravikumar and Lafferty, 2006 SOCP-MS ≡ QP-RL

LP-S vs. SOCP-MS Differ in the way they relax X = xx T X ij  [-1,1] 1 + x i + x j + X ij ≥ 0 ∑ X ij = (2 - |L|) x i j  V b LP-S (x i + x j ) 2  2 + 2X ij (x i - x j ) 2  2 - 2X ij SOCP-MS F(LP-S) F(SOCP-MS)

LP-S vs. SOCP-MS LP-S strictly dominates SOCP-MS LP-S strictly dominates QP-RL Where have we gone wrong? A Quick Recap !

Recap of SOCP-MS xixi xjxj X ij Can we use different C matrices ?? Can we use a different subgraph ?? 1 1 C = (x i - x j ) 2  2 - 2X ij C = (x i + x j ) 2  2 + 2X ij

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results –SOCP Relaxations on Trees –SOCP Relaxations on Cycles Two New SOCP Relaxations

SOCP Relaxations on Trees Choose any arbitrary tree

SOCP Relaxations on Trees Choose any arbitrary C 0 Repeat over trees to get relaxation SOCP-T LP-S strictly dominates SOCP-T LP-S strictly dominates QP-T

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results –SOCP Relaxations on Trees –SOCP Relaxations on Cycles Two New SOCP Relaxations

SOCP Relaxations on Cycles Choose an arbitrary even cycle P ij ≥ 0 P ij ≤ 0OR

SOCP Relaxations on Cycles Choose any arbitrary C 0 Repeat over even cycles to get relaxation SOCP-E LP-S strictly dominates SOCP-E LP-S strictly dominates QP-E

SOCP Relaxations on Cycles True for odd cycles with P ij ≤ 0 True for odd cycles with P ij ≤ 0 for only one edge True for odd cycles with P ij ≥ 0 for only one edge True for all combinations of above cases

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results Two New SOCP Relaxations –The SOCP-C Relaxation –The SOCP-Q Relaxation

The SOCP-C Relaxation Include all LP-S constraintsTrue SOCP ab cd Cycle of size 4 Define SOCP Constraint using appropriate C SOCP-C strictly dominates LP-S Which SOCP is strictly dominated by cycle inequalities? Open Question !!!

Outline Integer Programming Formulation Existing Relaxations Comparison Generalization of Results Two New SOCP Relaxations –The SOCP-C Relaxation –The SOCP-Q Relaxation

The SOCP-Q Relaxation Include all cycle inequalitiesTrue SOCP ab cd Clique of size n Define an SOCP Constraint using C = 1 SOCP-Q strictly dominates LP-S SOCP-Q strictly dominates cycle inequalities

Conclusions Large class of SOCP/QP dominated by LP-S New SOCP relaxations dominate LP-S Preliminary experiments conform with analysis

Future Work Comparison with cycle inequalities Determine best SOC constraints Develop efficient algorithms for new relaxations

Part 2

4-Neighbourhood MRF Test SOCP-C 50 binary MRFs of size 30x30 u ≈ N (0,1) P ≈ N (0,σ 2 )

4-Neighbourhood MRF σ = 2.5

8-Neighbourhood MRF Test SOCP-Q 50 binary MRFs of size 30x30 u ≈ N (0,1) P ≈ N (0,σ 2 )

8-Neighbourhood MRF σ = 1.125

Equivalent Relaxations A dominates B A B = B dominates A For all MAP Estimation problem (u, P)

Strictly Dominating Relaxation A dominates B A B > B does not dominate A For at least one MAP Estimation problem (u, P)