1 Game Theory Sequential bargaining and Repeated Games Univ. Prof.dr. M.C.W. Janssen University of Vienna Winter semester 2010-11 Week 46 (November 14-15)

Slides:



Advertisements
Similar presentations
Vincent Conitzer CPS Repeated games Vincent Conitzer
Advertisements

Some Problems from Chapt 13
Infinitely Repeated Games
Crime, Punishment, and Forgiveness
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Game Theory “Доверяй, Но Проверяй” (“Trust, but Verify”) - Russian Proverb (Ronald Reagan) Topic 5 Repeated Games.
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
Non-Cooperative Game Theory To define a game, you need to know three things: –The set of players –The strategy sets of the players (i.e., the actions they.
Chapter 14 Infinite Horizon 1.Markov Games 2.Markov Solutions 3.Infinite Horizon Repeated Games 4.Trigger Strategy Solutions 5.Investing in Strategic Capital.
M9302 Mathematical Models in Economics Instructor: Georgi Burlakov 2.5.Repeated Games Lecture
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
Infinitely Repeated Games Econ 171. Finitely Repeated Game Take any game play it, then play it again, for a specified number of times. The game that is.
Repeated games with Costly Observations Eilon Solan, Tel Aviv University Ehud Lehrer Tel Aviv University with.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
Game Theory Lecture 9.
Industrial Organization - Matilde Machado Tacit Collusion Tacit Collusion Matilde Machado.
Game-theoretic analysis tools Necessary for building nonmanipulable automated negotiation systems.
Game Theory Lecture 8.
Games People Play. 8: The Prisoners’ Dilemma and repeated games In this section we shall learn How repeated play of a game opens up many new strategic.
Dynamic Games of Complete Information.. Repeated games Best understood class of dynamic games Past play cannot influence feasible actions or payoff functions.
EC941 - Game Theory Prof. Francesco Squintani Lecture 8 1.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Rational Learning Leads to Nash Equilibrium Ehud Kalai and Ehud Lehrer Econometrica, Vol. 61 No. 5 (Sep 1993), Presented by Vincent Mak
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
ECON6036 1st semester Format of final exam Same as the mid term
APEC 8205: Applied Game Theory Fall 2007
Repeated games - example This stage game is played 2 times Any SPNE where players behave differently than in a 1-time game? Player 2 LMR L1, 10, 05, 0.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Static Games of Complete Information: Subgame Perfection
Extensive Game with Imperfect Information Part I: Strategy and Nash equilibrium.
Game Applications Chapter 29. Nash Equilibrium In any Nash equilibrium (NE) each player chooses a “best” response to the choices made by all of the other.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
© 2009 Institute of Information Management National Chiao Tung University Lecture Notes II-2 Dynamic Games of Complete Information Extensive Form Representation.
1. problem set 6 from Osborne’s Introd. To G.T. p.210 Ex p.234 Ex p.337 Ex. 26,27 from Binmore’s Fun and Games.
Game Theory.
MAKING COMPLEX DEClSlONS
Punishment and Forgiveness in Repeated Games. A review of present values.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Dynamic Games & The Extensive Form
Moshe Tennenholtz, Aviv Zohar Learning Equilibria in Repeated Congestion Games.
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Bargaining Theory MIT Game Theory. Bargaining Theory Cooperative (Axiomatic) –Edgeworth –Nash Bargaining –Variations of Nash –Shapley Value Non-cooperative.
Lecture 5 Leadership and Reputation Reputations arise in situations where there is an element of repetition, and also where coordination between players.
제 10 장 게임이론 Game Theory: Inside Oligopoly
Punishment, Detection, and Forgiveness in Repeated Games.
Final Lecture. Problem 2, Chapter 13 Exploring the problem Note that c, x yields the highest total payoff of 7 for each player. Is this a Nash equilibrium?
Mixed Strategies and Repeated Games
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
Lecture 6 Reputation Reputations arise in situations where there is an element of repetition, and also where coordination between players is possible.
Game Theory (Microeconomic Theory (IV)) Instructor: Yongqin Wang School of Economics, Fudan University December, 2004.
Lec 23 Chapter 28 Game Theory.
Entry Deterrence Players Two firms, entrant and incumbent Order of play Entrant decides to enter or stay out. If entrant enters, incumbent decides to fight.
Nash Bargaining Solution and Alternating Offer Games MIT Game Theory.
Lecture V: Bargaining Recommended Reading: Dixit & Skeath, Chapter 17 Osborne, Chapter 6.1, 16 Powell, In the Shadow of Power, Ch. 3.
Yuan Deng Vincent Conitzer Duke University
ECON 330 Lecture 17 Monday, November 25.
Dynamic Games of Complete Information
Vincent Conitzer CPS Repeated games Vincent Conitzer
Multiagent Systems Repeated Games © Manfred Huber 2018.
Vincent Conitzer Repeated games Vincent Conitzer
Chapter 14 & 15 Repeated Games.
Chapter 14 & 15 Repeated Games.
Molly W. Dahl Georgetown University Econ 101 – Spring 2009
Vincent Conitzer CPS Repeated games Vincent Conitzer
Presentation transcript:

1 Game Theory Sequential bargaining and Repeated Games Univ. Prof.dr. M.C.W. Janssen University of Vienna Winter semester Week 46 (November 14-15)

2 Sequential Bargaining Ultimatum game is a sequential bargaining game with one round.  SPE we know Consider then a sequential bargaining game with two rounds and alternating offers, and players discounting future pay-off with δ.  SPE pay-offs are (1-δ, δ)  Player 2 can propose to keep everything in last round and this will be accepted. Thus, by refusing in the first round he can guarantee himself δ  Player 1 should give him at least δ in first round if 2 is about to accept; he can get at most 1-δ

3 Alternating offers (Rubinstein, Stahl) Same stage game, but repeated infinitely often. What are equilibrium profits?  Define v (v*) as lowest (highest) pay-off you can get if you make an offer  Because of infinite horizon and equal discount factors, period 1 analysis is the same as period 2 analysis  v ≥1- δv*: lowest pay-off player 1 can guarantee himself is remaining of highest discounted pay-off player 2 can guarantee himself in the next round  v* ≤ 1- δv : highest pay-off player 1 can guarantee himself is remaining of lowest discounted pay-off player 2 can guarantee himself in the next round  v ≥1/(1+ δ) and v* ≤ 1/(1+ δ). Hence, equalities have to hold  Player 1 is better off as he makes first proposal, but advantage disappears when δ gets close to 1. Intuitive  First offer such that it is immediately accepted! Why to bother about rest of the game?  Unique subgame perfect equilibrium strategies

4 What if δ’s differ across players Period 1 analysis is similar to period 3 analysis, but not anymore to period 2 analysis  Define v i (v i *) as lowest (highest) pay-off player i can get if she makes an offer  v 1 ≥ 1- δ 2 v 2 *; by symmetry, the same thing holds for player 2.  v 1 * ≤ 1- δ 2 v 2 ; by symmetry, the same thing holds for player 2.  v 1 ≥ (1- δ 2 )/(1- δ 1 δ 2 ) and v* 1 ≤ (1- δ 2 )/(1- δ 1 δ 2 )  Hence, equalities have to hold; additional advantage for player with highest δ.

5 Notation in repeated games Define history of play as follows. Let a 0 = (a 0 1, a 0 2,…,a 0 n ) be the action profile that is played in stage 0, i.e., the actions played by all players History at the beginning of period 1, h 1 = a 0 History at the beginning of stage t+1, h t+1 = (a 0,…,a t ) The set H t is the set of all possible histories h t and A i (h t ) is the set of actions that player i can choose after history h t and A i (H t ) is the union of this set over all possible histories Strategy σ i of player i is a sequence of mappings {σ k i } where each σ k i maps H k to mixed actions.  Note that you cannot condition on the random events

6 Subgame perfection and the one-stage deviation principle in finitely repeated games One stage deviation principle: No player can deviate by deviating in a single period and then returning back to the (equilibrium) strategy  There is no player i and strategy s’(i) that is equal to s*(i) apart from the action in one period given one history h, such that u i (s’ i,s* -i ) > u i (s* i,s* -i ) given that history h Prop. In finite horizon games, a strategy combination s* is a SPE if, and only if, it satisfies “one stage deviation principle”.  Only if: clear, otherwise there is an immediate violation of SPE definition  If: suppose to the contrary, s* satisfies the principle but is not SPE. Then there is a stage t and a history h t s.t. at least one player has a strategy s’ i (h t )≠s* i (h t ) and s’ i (h t ) is a better response. Continuation next slide

7 Proof one stage deviation principle Let t’ be the last period in which s’ i (h t’ )≠s* i (h t’ ) Because of the one-stage-deviation principle t’ > t Period t’ is defined such that for all t” > t’ s’ i (h t” )=s* i (h t” ) Define then another strategy s I that is such that it coincides with s’ I up to t’ and coincides with s* I at t’ and afterwards. Because of the one-stage-deviation principle and since s’ i (h t” )=s i (h t” ) for all t” > t’, s i is as good a response given history h t If t’ = t+1, then s i only differs in one period from s*, and therefore the one stage deviation principle implies that s i cannot be strictly better If t’ > t+1, similar argument applies (details page 109)

8 Additional equilibria in repeated games Main interest in repeated games is what type of equilibrium outcomes can be supported that cannot be supported in a static game  Repetition of static equilibrium is always an equilibrium in a repeated game; not so interesting  Thus, what else?  Consider an example

9 A Static Game

10 Multiple Equilibria Nash Equilibria

11 Can non-Nash outcomes of the static game be supported in equilibrium if the game is repeated 2 times?

12 Last period analysis In the last period they cannot choose for (U,L) As both firms have an incentive to “cheat” as 16 is a higher pay-off than 12 Punishment is not possible (as it is the last period)

13 First-period analysis But: in the first period they can choose for (U,L) Strategy: - Choose “U (L) ” in period 1 - Choose “M (C)” in period 2 when other chooses “L (U)” in period 1 -Choose “B (R)” in period 2 when other chooses somthing else in period 1 Punishment is part of strategy Is this an equilibrium? Is it a SPE

14 Pay-offs in infinitely repeated games Overall pay-offs u i ; stage game pay-offs gi, continuation pay-off from period t onwards Want to have an expression where one can easily compare stage game pay-offs and repeated game pay-offs, i.e., normalisation: Time averaging is sometimes used for the case of complete patience

15 Folk Theorem I If players are sufficiently patient, then any feasible, individually rational pay-offs can be enforced by an equilibrium  Individually rational pay-offs: minimax pay-off  v i =  m j i is action player j chooses to minimax player i  Feasible pay-offs is the convex hull V of the static game pay-offs, i.e., V = convex hull {v / there is an a  A such that g(a) =v}  Both terms need some explanation

16 Minimax pay-offs What are the Nash equilibria of this game?  Denote by q the probability player 2 chooses L  In a mixed strategy eq ⅓≤q≤⅔, pay-offs 0 and 1 Minimax for player 1  u(U) = -3q+1  u(M) = 3q-2  U(D) = 0  Minimax is 0 Minimax for player 2 is also 0  By 1 choosing (½,½,0) Thus, minimax pay-offs can be lower than Nash eq. pay-offs LR U-2,21,-2 M -2,2 D0,1

17 Feasible pay-offs Equilibrium pay-offs are (2,1), (1,2) and (⅔, ⅔) Convex hull of eq. pay-offs is triangle connecting the three points (also e.g. (1½,1½)) V connects (2,1), (1,2) and (0, 0) But (1½,1½) cannot be obtained by independent mixing, only as correlated eq Correlated mixing can happen in repeated setting by alternating between playing two equilibria (and time averaging pay-offs or δ close to 1) FB B0,02,1 F1,20,0 Eq. pay- offs

18 Folk Theorem II Prop. For every feasible pay-off vector v with v i > v i, there exist a δ δ there exist a Nash equilibrium of the infinitely repeated game with pay-off v.  Pay-offs in repeated game cannot only be larger, but also smaller than static Nash eq pay-offs!!  Basic idea: if players are sufficiently patient, then any finite gain in a one period deviation is nothing compared to a small, but permanent loss in future pay-offs (punishment by minimaxing a player)

19 “Proof” “Nash Folk Theorem” Consider feasible pay-off v and action profile g(a)=v  If there is no action profile a that yields v, you may choose a sequence of actions such that v is (close to) average (discounted) pay-offs (or a public randomization) Consider strategy: start by playing a i ; play a i as long as others do, if one player j deviates minimax him forever, i.e., choose m j i Deviation in period t yields normalised pay-off which is smaller than v i if δ is larger than δ i, where δ i solves

20 Is the threat of Minimaxing credible? If we restrict analysis to static “Nash threats”, then Friedman shows that only pay-offs larger than the static Nash equilibrium pay- offs can be supported Others show that in games where the minimax pay-offs are lower than the static equilibrium pay-offs, even worse outcomes can be compatible with a SPE of the infinitely repeated game.

21 Basic idea of SPE with minimax pay-offs time averaging After a deviation, play the minimax pay-off for N periods, where N is chosen for all players s.t. After N periods return back to “cooperative” mood (finite) N ensures that no player has an incentive to deviate Cost of punishment is extremely small as with time averaging pay-offs in a finite number of periods “do not make a difference”  Average pay-off to player j when i is punished is v j

22 Basic idea of SPE with minimax pay-offs discounted pay-offs Previous strategies (for time averaging pay-offs) do not work as it may be that minimaxing another player gives a player a lower pay-off than his own minimax pay-off. Reward punishers, instead of punishing them if they don’t punish Choose a vector in the interior of V such that for each i you can still give a higher pay-off.  V needs to be of “full dimension” Play in three phases:  Initial cooperative phase  Punishment phase where players minimax for N(j) periods the deviator j (as before); switch to punishment phase for player i if i deviates in one of the N(j) periods.  Reward phase after the punishment phase is fully completed

23 Renegotiation proofness in repeated games Is SPE the best notion of a credible threat? Suppose you cooperate for some time in the PD and then someone defects, by chance. Should you go back immediately to always defect?  Or should players “renegotiate”?  It is in both players interest to revert back to the cooperative outcome  In any subgame the equilibrium played must not be Pareto- dominated.  Pareto-optimality as an assumption and the critique that is possible (risk dominance and Pareto-dominance)  Deviations are accidents and unlikely to be repated? “Bygones are bygones”

24 Pareto perfection only applies in two- player games Two Nash equilibria in pure strategies: (U,L,A) and (D,R,B) ULA is Pareto-efficient Natural candidate? Suppose players 1 and 2 expect matrix chooser to choose A. Then they can renegotiate and gain by playing (D,R) LR U 0,0,10-5,-5,0 D 1,1,-5 A LR U -2,-2,0-5,-5,0 D -1,-1,5 B

25 Definition of Pareto perfect equilibrium Fix stage game g and play it for T periods. Let P(T) the set of pay-offs of pure strategy SPE of G(T) R(t) is the set of strongly efficient points of P(t), i.e., this is the set of points such that there is not another pay-off point where no player is worse off and some player is better off. Set Q(1) = P(1) For any t, let Q(t) be the set of pay-offs of pure strategy SPE that can be enforced with continuation pay-offs in R(t-1) A SPE is Pareto perfect if for every possible history and in every time period t, the continuation pay-offs are in R(T-t)

26 Pareto perfection restricts threats Some efficient equilibria cannot be supported anymore under Pareto-perfection It restricts the set of threats, and thereby it is more difficult to keep players on the equilibrium path Example

27 Example Pareto-perfection c1c2c3c4 R10,01,40,06,0 R24,10,0 R30,0 3,30,0 R40,60,0 5,5 Three pure strategies in G(1) with pay-offs (4,1), (1,4) and (3,3) In G(2) without discounting pay- off of 8 is possible. Unique element in R(2) Without restriction to Pareto perfection in G(3) pay-off of 13 possible With Pareto perfection in first period of G(3) no threat possible; one has to play stage game equilibrium Equilibrium play alternates between odd and even periods under Pareto perfection