The Duality Theorem Primal P: Maximize

Slides:



Advertisements
Similar presentations
Thursday, March 7 Duality 2 – The dual problem, in general – illustrating duality with 2-person 0-sum game theory Handouts: Lecture Notes.
Advertisements

Tuesday, March 5 Duality – The art of obtaining bounds – weak and strong duality Handouts: Lecture Notes.
Duality for linear programming. Illustration of the notion Consider an enterprise producing r items: f k = demand for the item k =1,…, r using s components:
1 LP Duality Lecture 13: Feb Min-Max Theorems In bipartite graph, Maximum matching = Minimum Vertex Cover In every graph, Maximum Flow = Minimum.
Geometry and Theory of LP Standard (Inequality) Primal Problem: Dual Problem:
Totally Unimodular Matrices
C&O 355 Mathematical Programming Fall 2010 Lecture 12 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Introduction to Algorithms
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
Game theory.
Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc
9.4 Linear programming and m x n Games: Simplex Method and the Dual Problem In this section, the process of solving 2 x 2 matrix games will be generalized.
Part 3: The Minimax Theorem
Basic Feasible Solutions: Recap MS&E 211. WILL FOLLOW A CELEBRATED INTELLECTUAL TEACHING TRADITION.
CS38 Introduction to Algorithms Lecture 15 May 20, CS38 Lecture 15.
Duality Lecture 10: Feb 9. Min-Max theorems In bipartite graph, Maximum matching = Minimum Vertex Cover In every graph, Maximum Flow = Minimum Cut Both.
Duality Dual problem Duality Theorem Complementary Slackness
An Introduction to Game Theory Part III: Strictly Competitive Games Bernhard Nebel.
Computational Methods for Management and Economics Carla Gomes
7(2) THE DUAL THEOREMS Primal ProblemDual Problem b is not assumed to be non-negative.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 68 Chapter 9 The Theory of Games.
Minimax strategies, Nash equilibria, correlated equilibria Vincent Conitzer
Use complementary slackness to check the solution: ( 20/3, 0, 16/3, 0) Maximize 9 x 1 -3 x x 3 -7 x 4 subject to 2 x x x x 4 ≤
Duality Theory 對偶理論.
Duality Theory LI Xiaolei.
Duality Theory  Every LP problem (called the ‘Primal’) has associated with another problem called the ‘Dual’.  The ‘Dual’ problem is an LP defined directly.
TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A A A A A A Image:
Advanced Operations Research Models Instructor: Dr. A. Seifi Teaching Assistant: Golbarg Kazemi 1.
 Minimization Problem  First Approach  Introduce the basis variable  To solve minimization problem we simple reverse the rule that is we select the.
1 Bob and Sue solved this by hand: Maximize x x 2 subject to 1 x x 2 ≤ x x 2 ≤ 4 x 1, x 2 ≥ 0 and their last dictionary was: X1.
Part 3 Linear Programming 3.3 Theoretical Analysis.
Linear Program Set Cover. Given a universe U of n elements, a collection of subsets of U, S = {S 1,…, S k }, and a cost function c: S → Q +. Find a minimum.
Hon Wai Leong, NUS (CS6234, Spring 2009) Page 1 Copyright © 2009 by Leong Hon Wai CS6234: Lecture 4  Linear Programming  LP and Simplex Algorithm [PS82]-Ch2.
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
Copyright © 2006 Brooks/Cole, a division of Thomson Learning, Inc. Linear Programming: An Algebraic Approach 4 The Simplex Method with Standard Maximization.
9.2 Mixed Strategy Games In this section, we look at non-strictly determined games. For these type of games the payoff matrix has no saddle points.
Approximation Algorithms Duality My T. UF.
2.5 The Fundamental Theorem of Game Theory For any 2-person zero-sum game there exists a pair (x*,y*) in S  T such that min {x*V. j : j=1,...,n} =
Approximation Algorithms based on linear programming.
Game Theory [geym theer-ee] : a mathematical theory that deals with the general features of competitive situations in a formal abstract way.
Mixed Strategies Keep ‘em guessing.
EMGT 6412/MATH 6665 Mathematical Programming Spring 2016
Lap Chi Lau we will only use slides 4 to 19
Integer Programming An integer linear program (ILP) is defined exactly as a linear program except that values of variables in a feasible solution have.
Chap 10. Sensitivity Analysis
Game Theory Just last week:
Tools for Decision Analysis: Analysis of Risky Decisions
Topics in Algorithms Lap Chi Lau.
The minimum cost flow problem
EMGT 6412/MATH 6665 Mathematical Programming Spring 2016
Perturbation method, lexicographic method
Duality for linear programming.
Game Theory.
Chap 9. General LP problems: Duality and Infeasibility
The Simplex Method: Standard Minimization Problems
Analysis of Algorithms
Duality Theory and Sensitivity Analysis
Part 3 Linear Programming
Chapter 4. Duality Theory
Chapter 8. General LP Problems
Chapter 5. The Duality Theorem
Lecture 20 Linear Program Duality
Flow Feasibility Problems
9.3 Linear programming and 2 x 2 games : A geometric approach
Chapter 8. General LP Problems
DUALITY THEORY Reference: Chapter 6 in Bazaraa, Jarvis and Sherali.
Chapter 8. General LP Problems
Chapter 2. Simplex method
Presentation transcript:

The Duality Theorem Primal P: Maximize 𝑐 ⊺ 𝑥 subject to 𝐴𝑥≤𝑏, 𝑥≥0. Dual D: Minimize 𝑏 ⊺ 𝑦 subject to 𝐴 ⊺ 𝑦≥𝑐, 𝑦≥0. Weak Duality Theorem: If 𝑥 is feasible for P and 𝑦 is feasible for D, then: 𝑐 ⊺ 𝑥≤ 𝑏 ⊺ 𝑦 (Strong) Dualiy Theorem: If 𝑥 ∗ is optimal for P and 𝑦 ∗ is optimal for D, then: 𝑐 ⊺ 𝑥 ∗ = 𝑏 ⊺ 𝑦 ∗

Generalized Duality Maximize Subject to Primal: Minimize Subject to 𝑐 1 𝑥 1 + 𝑐 2 𝑥 2 + 𝑐 3 𝑥 3 Subject to 𝑦 1 : 𝑎 11 𝑥 1 + 𝑎 12 𝑥 2 + 𝑎 13 𝑥 3 ≤ 𝑏 1 𝑦 2 : 𝑎 21 𝑥 1 + 𝑎 22 𝑥 2 + 𝑎 23 𝑥 3 ≥ 𝑏 2 𝑦 3 : 𝑎 31 𝑥 1 + 𝑎 32 𝑥 2 + 𝑎 33 𝑥 3 = 𝑏 3 𝑥 1 ≥0, 𝑥 2 ≤0, 𝑥 3 free Primal: Minimize 𝑏 1 𝑦 1 + 𝑏 2 𝑦 2 + 𝑏 3 𝑦 3 Subject to 𝑥 1 : 𝑎 11 𝑦 1 + 𝑎 21 𝑦 2 + 𝑎 31 𝑦 3 ≥ 𝑐 1 𝑥 2 : 𝑎 12 𝑦 1 + 𝑎 22 𝑦 2 + 𝑎 32 𝑦 3 ≤ 𝑐 2 𝑥 3 : 𝑎 13 𝑦 1 + 𝑎 23 𝑦 2 + 𝑎 33 𝑦 3 = 𝑐 3 𝑦 1 ≥0, 𝑦 2 ≤0, 𝑦 3 free Dual:

Rules for Taking the Dual Constraints in primal corresponds to variables in dual (and vice versa). Coefficients of objective function in primal corresponds to constants in dual (and vice versa). Primal (Maximization) Dual (Minimization) ≤ for constraint ≥0 for variable ≥ for constraint ≤0 for variable = for constraint free for variable

Certifying optimality Suppose we wish to prove to someone that a solution 𝑥 ∗ is optimal. If we supply both the optimal solution 𝑥 ∗ and an optimal solution 𝑦 ∗ to the dual, then one may easily verify that 𝑥 ∗ is in fact optimal. Can we similarly prove a linear program is infeasible or unbounded?

Certifying infeasibility Farkas Lemma: Exactly one of the following is true: There exist 𝑥 s.t. 𝐴𝑥≤𝑏. There exist 𝑦≥0 s.t. 𝐴 ⊤ 𝑦=0 and 𝑏 ⊤ 𝑦<0. Proof: Consider the following program and its dual P: max 0 s.t. 𝐴𝑥≤𝑏 D: min 𝑏 ⊤ 𝑦 s.t. 𝐴 ⊤ 𝑦=0, 𝑦≥0

Proof of Farkas Lemma P: max 0 s.t. 𝐴𝑥≤𝑏 D: min 𝑏 ⊤ 𝑦 s.t. 𝐴 ⊤ 𝑦=0, 𝑦≥0 Assume P is infeasible. Then D is either infeasible or unbounded. D is always feasible, hence unbounded. Thus we may find feasible solution 𝑦 with 𝑏 ⊤ 𝑦<0.

The Complementary Slackness Property Primal P: Maximize 𝑐 ⊺ 𝑥 subject to 𝐴𝑥≤𝑏, 𝑥≥0. Dual D: Minimize 𝑏 ⊺ 𝑦 subject to 𝐴 ⊺ 𝑦≥𝑐, 𝑦≥0. Solutions 𝑥 and 𝑦 are said to satisfy Complementary Slackness if and only if: for all 𝑖=1,2,…,𝑚 (at least) one of the following holds: 𝑗=1 𝑛 𝑎 𝑖𝑗 𝑥 𝑗 = 𝑏 𝑖 (𝑖th primal constraint has slack 0) 𝑦 𝑖 =0 (𝑖th dual variable is 0) for all j=1,2,…,𝑛 (at least) one of the following holds: 𝑖=1 𝑚 𝑎 𝑖𝑗 𝑦 𝑖 = 𝑐 𝑗 (𝑗th dual constraint has slack 0) 𝑥 𝑗 =0 (𝑗th primal variable is 0)

Complementary Slackness Theorem Primal P: Maximize 𝑐 ⊺ 𝑥 subject to 𝐴𝑥≤𝑏, 𝑥≥0. Dual D: Minimize 𝑏 ⊺ 𝑦 subject to 𝐴 ⊺ 𝑦≥𝑐, 𝑦≥0. Theorem Let 𝑥 and 𝑦 be feasible solutions to P and D. Then: 𝑥 and 𝑦 are both optimal if and only if 𝑥 and 𝑦 satisfy complementary slackness.

Strict Complementary Slackness Property Primal P: Maximize 𝑐 ⊺ 𝑥 subject to 𝐴𝑥≤𝑏, 𝑥≥0. Dual D: Minimize 𝑏 ⊺ 𝑦 subject to 𝐴 ⊺ 𝑦≥𝑐, 𝑦≥0. Solutions 𝑥 and 𝑦 are said to satisfy Strict Complementary Slackness if and only if: for all 𝑖=1,2,…,𝑚 exactly one of the following holds: 𝑗=1 𝑛 𝑎 𝑖𝑗 𝑥 𝑗 = 𝑏 𝑖 (𝑖th primal constraint has slack 0) 𝑦 𝑖 =0 (𝑖th dual variable is 0) for all j=1,2,…,𝑛 exactly one of the following holds: 𝑖=1 𝑚 𝑎 𝑖𝑗 𝑦 𝑖 = 𝑐 𝑗 (𝑗th dual constraint has slack 0) 𝑥 𝑗 =0 (𝑗th primal variable is 0)

Strict Complementary Slackness Theorem If P has an optimal solution then there exist optimal solution 𝑥 to P and 𝑦 to D that satisfy strict complementary slackness. Proof: Let 𝑧 ∗ be value of an optimal solution. Suppose 𝑗 is such that 𝑥 𝑗 =0 in all optimal solutions. Then P’: max x j s.t. 𝐴𝑥≤𝑏, −𝑐 ⊤ 𝑥≤− 𝑧 ∗ , 𝑥≥0 has optimal value 0. D’: min 𝑏 ⊤ 𝑦− 𝑧 ∗ 𝑡 s.t. 𝐴 ⊤ 𝑦−𝑐𝑡≥ 𝑒 𝑗 ,𝑦≥0,𝑡≥0

Proof (cont’d) If 𝑥 𝑗 =0 in all optimal solutions, then D’: min 𝑏 ⊤ 𝑦− 𝑧 ∗ 𝑡 s.t. 𝐴 ⊤ 𝑦−𝑐𝑡≥ 𝑒 𝑗 ,𝑦≥0,𝑡≥0 has optimal value 0. Let ( 𝑦 ∗ , 𝑡 ∗ ) be any optimal solution to D’. Case 𝑡 ∗ =0: Then we have 𝑏 ⊤ 𝑦 ∗ =0 𝐴 ⊤ 𝑦 ∗ ≥ 𝑒 𝑗 𝑦 ∗ ≥0 Let 𝑦 be optimal solution to D. Then 𝑦+ 𝑦 ∗ is also optimal solution of D, and has nonzero slack in 𝑗th constraint.

Proof (cont’d) Case 𝑡 ∗ >0: Then we have 𝑏 ⊤ 𝑦 ∗ − 𝑧 ∗ 𝑡=0 𝐴 ⊤ 𝑦 ∗ −𝑐𝑡≥ 𝑒 𝑗 𝑦 ∗ , 𝑡 ∗ ≥0 Define 𝑦= 𝑦 ∗ / 𝑡 ∗ . Then we have 𝑏 ⊤ 𝑦= 𝑧 ∗ 𝐴 ⊤ 𝑦≥𝑐+ 𝑒 𝑗 / 𝑡 ∗ 𝑦 ∗ ≥0 Hence 𝑦 is optimal solution to D and has nonzero slack in 𝑗th constraint.

Proof (cont’d) For each 𝑗 where all optimal solutions to P satisfies 𝑥 𝑗 =0 we have found an optimal solution 𝑦 (𝑗) to D with nonzero slack in 𝑗th constraint. The average of all these 𝑦 (𝑗) is an optimal solution to D with nonzero slack in 𝑗th constraint for all 𝑗 we have considered. Finally do the same for the dual finding solution 𝑥. Then 𝑥 and 𝑦 satisfy strict complementary slackness.

Game Theory Just last week: CMU poker AI player Libratus beats top human poker players in heads up no-limit Texas Hold’em. A monumental achievement! (Compare to Chess (1997), Jeopardy (2009), Go (2016)).

Guess the coin game Minnie hides in an envelope some money. It is either a 1 DKK coin, a 5 DKK coin, or a 10 DKK coin. If Max can guess what it is, he gets the coin!

How should Max play? - Worst case approach Max does not know Minnie and has no clue as to what coin she has hidden. He wants to find a strategy with a good worst case guarantee. He cannot get a worst case guarantee on his actual winnings. But he can get a non-trivial worst case guarantee on his expected winnings, if he uses a randomized strategy!

Guess the coin game - Minnie's perspective Somehow Minnie has been convinced to play the game with Max. But she does actually not like to just give her money away. Can we find the randomized strategy for Minnie that minimizes her expected loss?

Matrix Games A Matrix Game is given by a payoff matrix 𝐴 =( 𝑎 𝑖𝑗 )∈ 𝑅 𝑚×𝑛 . Row player (Minnie) chooses row 𝑖∈ 1,…,𝑚 (without seeing Max's move). Column player (Max) chooses column 𝑗∈{1,…𝑛} (without seeing Minnies's move). Max gains 𝑎 𝑖𝑗 “dollars” and Minnie loses 𝑎 𝑖𝑗 “dollars”. Matrix games are zero-sum games as Max gains exactly what Minnie loses. Negative numbers can be used to model money transferred in the other direction. Warning: In many texts and old exam problems, the Row player is the maximizer and the Column player is the minimizer.

Guess the coin as a matrix game Guess DKK 1 Guess DKK 5 Guess DKK 10 Hide DKK 1 1 Hide DKK 5 5 Hide DKK 10 10

Optimal randomized Strategy (for Max) Play the game in a randomized way so that the expected gain is as big as possible, assuming worst case behavior of Minnie.

Column Players (Max's) Optimal randomized Strategy Optimal randomized strategy and guaranteed lower bound on exp. gain for Max is ( 𝑝 1 , 𝑝 2 ,…, 𝑝 𝑛 ;𝑔) which is a solution to the LP: max 𝑔 s.t. 𝑗=1 𝑛 𝑎 𝑖𝑗 𝑝 𝑗 ≥𝑔 𝑖=1,…,𝑚 𝑗=1 𝑛 𝑝 𝑗 =1 𝑝 𝑗 ≥0 𝑗=1,…,𝑛

Row Players (Minnie's) Optimal randomized Strategy Optimal randomized strategy and guaranteed upper bound on exp. loss for Minnie is ( 𝑞 1 , 𝑞 2 ,…, 𝑞 𝑚 ;𝑙) which is a solution to the LP: min 𝑙 s.t. 𝑖=1 𝑚 𝑎 𝑖𝑗 𝑞 𝑖 ≤𝑙 j=1,…,𝑛 𝑖=1 𝑚 𝑞 𝑖 =1 𝑞 𝑖 ≥0 i=1,…,𝑚

Crucial observation Max's program and Minnie's program are each others duals!

Consequence For any Matrix game, Max's guaranteed lower bound on his expected gain when he plays his optimal randomized strategy is equal to Minnie's guaranteed upper bound on her expected loss when she plays her optimal randomized strategy. The common value is thus called the value of the game. A priori, this is not obvious - intuitively, the “optimal" strategies are very timid and “pessimistic", as they are optimal only when assuming a worst case opponent.

"Guess the basis" trick for solving linear programs and matrix games If we know the basis of a basic solution to an LP in standard form, we can find it directly by solving the system of linear equations setting the non-basic variables to zero. If we can guess the basis of an optimal solution to the primal and guess the basis of an optimal solution to the dual, we can therefore find the corresponding basic solutions by solving two systems of linear equations, and verify that both are optimal by checking that there is no gap between solution values.

"Guess the basis" trick for solving linear programs and matrix games “knowing the basis” = “knowing which variables are strictly positive” (unless the LP is degenerate) “Default" guess for the bases of the LPs corresponding to a square matrix game: The probabilities of playing the row/columns are all strictly positive (no strategies are "stupid"), so the variables that should be set to zero are the slack variables of all inequalities.

Fair and symmetric games A matrix game is called symmetric if its matrix is antisymmetric(!). Example: Rock/Scissors/Paper. Counterexample: Guess the coin. A matrix game is called fair if its value is 0. Theorem: All symmetric games are fair. The opposite is not necessarily true. Example?

Matrix algebra formulation Note: If Max plays by (not necessarily optimal) randomized strategy 𝑝= 𝑝 1 ,…, 𝑝 𝑛 and Minnie plays by the randomized strategy 𝑞= 𝑞 1 ,…, 𝑞 𝑚 the expected gain of Max is 𝑞 ⊤ 𝐴𝑝. If the value of the game is 𝑣, we have that 𝑣= max 𝑝 min 𝑞 𝑞 ⊤ 𝐴𝑝 Proof: 𝑣 is Max's best possible expected gain, assuming a worst case move of Minnie and hence also assuming a worst case randomized strategy of Minnie.

von Neuman Min-Max Theorem (Theorem 11.1) For any real matrix 𝐴, max 𝑝 min 𝑞 𝑞 ⊤ 𝐴𝑝 = min 𝑞 max 𝑝 𝑞 ⊤ 𝐴𝑝 where 𝑝 (resp. 𝑞) are arbitrary probability distributions on columns (resp. rows) of 𝐴.

Exploiting weak opponents Can we play an optimal strategy and still exploit opponents that do not play by the optimal strategy? Exploit: Achieve a better expectation that the value of the game.

Bad news: Principle of indifference Suppose 𝑝 ∗ is optimal for Max and 𝑞 ∗ is optimal for Minnie. Let 𝑆⊆{1,2,…,𝑚} be the set of rows to which 𝑞 ∗ assigns non-zero probability. Suppose Max plays 𝑝 ∗ . Then his expected payoff is the same, namely the value of the game, no matter which actual row in 𝑆 Minnie chooses. Proof: Complementary Slackness Theorem! If 𝑞 𝑖 ∗ >0 then 𝑖th constraint is satisfied with equality by 𝑝 ∗ .

Exploiting weak opponents - good news! Let 𝑆⊆{1,2,…,𝑚} be the set of rows to which some optimal strategy 𝑞 ∗ for Minnie assigns non-zero probability. There exists an optimal strategy 𝑝 ∗ for Max so that his expected payoff is strictly bigger than the value of the game if Minnie chooses a row that is not in 𝑆 ∗ . Proof: Strict Complementary Slackness Theorem!

Best replies and Nash equilibrium Let 𝑦 be a randomized strategy for Minnie. A best reply to 𝑦 is a randomized strategy for Max that maximizes his expected winnings assuming that Minnie plays 𝑦. A pair of randomized strategies 𝑥 ∗ and 𝑦 ∗ is called a Nash equilibrium if they are best responses to each other. Theorem: A pair of randomized strategies ( 𝑥 ∗ , 𝑦 ∗ ) for a matrix game is a Nash equilibrium if and only if they are both optimal.