On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.

Slides:



Advertisements
Similar presentations
Infinitely Repeated Games
Advertisements

Crime, Punishment, and Forgiveness
G AME THEORY MILJAN KNEŽEVIĆ FACULTY OF MATHEMATICS UNIVERSITY IN BELGRADE.
BASICS OF GAME THEORY. Recap Decision Theory vs. Game Theory Rationality Completeness Transitivity What’s in a game? Players Actions Outcomes Preferences.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
3. Basic Topics in Game Theory. Strategic Behavior in Business and Econ Outline 3.1 What is a Game ? The elements of a Game The Rules of the.
COMP 553: Algorithmic Game Theory Fall 2014 Yang Cai Lecture 21.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
1 Evolution & Economics No Evolutionary Stability in Repeated Games Played by Finite Automata K. Binmore & L. Samuelson J.E.T Automata.
Congestion Games with Player- Specific Payoff Functions Igal Milchtaich, Department of Mathematics, The Hebrew University of Jerusalem, 1993 Presentation.
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
Non-Cooperative Game Theory To define a game, you need to know three things: –The set of players –The strategy sets of the players (i.e., the actions they.
6-1 LECTURE 6: MULTIAGENT INTERACTIONS An Introduction to MultiAgent Systems
Chapter 14 Infinite Horizon 1.Markov Games 2.Markov Solutions 3.Infinite Horizon Repeated Games 4.Trigger Strategy Solutions 5.Investing in Strategic Capital.
The basics of Game Theory Understanding strategic behaviour.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Calibrated Learning and Correlated Equilibrium By: Dean Foster and Rakesh Vohra Presented by: Jason Sorensen.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
Infinitely Repeated Games Econ 171. Finitely Repeated Game Take any game play it, then play it again, for a specified number of times. The game that is.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
EKONOMSKA ANALIZA PRAVA. Game Theory Outline of the lecture: I. What is game theory? II. Elements of a game III. Normal (matrix) and Extensive (tree)
17. (A very brief) Introduction to game theory Varian, Chapters 28, 29.
Game Theory Lecture 8.
Repeated Prisoner’s Dilemma If the Prisoner’s Dilemma is repeated, cooperation can come from strategies including: “Grim Trigger” Strategy – one.
Dynamic Games of Complete Information.. Repeated games Best understood class of dynamic games Past play cannot influence feasible actions or payoff functions.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Algoritmi per Sistemi Distribuiti Strategici
Basics on Game Theory For Industrial Economics (According to Shy’s Plan)
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
ECON6036 1st semester Format of final exam Same as the mid term
Game Theory: Key Concepts Zero Sum Games Zero Sum Games Non – Zero Sum Games Non – Zero Sum Games Strategic Form Games  Lay out strategies Strategic Form.
APEC 8205: Applied Game Theory Fall 2007
Game Theory Here we study a method for thinking about oligopoly situations. As we consider some terminology, we will see the simultaneous move, one shot.
PRISONER’S DILEMMA By Ajul Shah, Hiten Morar, Pooja Hindocha, Amish Parekh & Daniel Castellino.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
DANSS Colloquium By Prof. Danny Dolev Presented by Rica Gonen
Constraints in Repeated Games. Rational Learning Leads to Nash Equilibrium …so what is rational learning? Kalai & Lehrer, 1993.
An introduction to game theory Today: The fundamentals of game theory, including Nash equilibrium.
Inefficiency of equilibria, and potential games Computational game theory Spring 2008 Michal Feldman.
Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie Mellon University.
Punishment and Forgiveness in Repeated Games. A review of present values.
Monetary Economics Game and Monetary Policymaking.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Dynamic Games & The Extensive Form
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Public Policy Analysis MPA 404 Lecture 24. Previous Lecture Graphical Analysis of Tariff and Quota Game Theory; The prisoner's Dilemma.
Section 2 – Ec1818 Jeremy Barofsky
Subgames and Credible Threats (with perfect information) Econ 171.
Mohsen Afsharchi Multiagent Interaction. What are Multiagent Systems?
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
Intermediate Microeconomics Game Theory. So far we have only studied situations that were not “strategic”. The optimal behavior of any given individual.
Game representations, game-theoretic solution concepts, and complexity Tuomas Sandholm Computer Science Department Carnegie Mellon University.
Introduction to Game Theory
Communication Complexity as a Lower Bound for Learning in Games
Vincent Conitzer CPS Repeated games Vincent Conitzer
LECTURE 6: MULTIAGENT INTERACTIONS
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Multiagent Systems Repeated Games © Manfred Huber 2018.
Game Theory Fall Mike Shor Topic 5.
Vincent Conitzer Repeated games Vincent Conitzer
Chapter 14 & 15 Repeated Games.
Chapter 14 & 15 Repeated Games.
Collaboration in Repeated Games
Game Theory Spring Mike Shor Topic 5.
Vincent Conitzer CPS Repeated games Vincent Conitzer
Presentation transcript:

On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis

Bounded Rationality In Economic Theory computation is usually not considered as a pricey resource. Therefore it sometimes predicts that rational agents will invest a large amount of computation for small payoffs. If we bound the ability of agents to compute, new solutions might appear.

The Prisoner’s Dilemma 3,30,4 4,01,1 I II C D CD The Only Nash Equilibrium is (D,D). D strictly dominates C therefore: Problem: This contradicts natural human behavior.

The n-round Prisoner’s Dilemma Play the game n times. Compute the average outcome. A strategy depends upon the history of the game. Lemma: The only equilibrium is {D,D} n. Proof: Playing D in the last round dominates any other strategy. Use a backwards induction argument.

DC CD CC DD Paretto Optimal Threat Point Individually Rational

Goal: Use bounded rationality in order to find an equilibrium which approximates the collaborating outcome. Find an equilibrium in which both players play ‘C’ in most of the rounds.

Main Idea Limit the computational powers of the players in such a way that a new equilibrium would emerge. The players would be too ‘dumb’ to count the number of rounds therefore would not be able to defect in the last round. Use the fact that the number of strategies is double exponential in the number of rounds.

Model Of Computation The model of computation is automata. Each pure strategy would be played by a single automaton. A mixed strategy is a distribution over automata The resource we limit is the number of states an automaton is aloud to have.

Automata Theory - Basics The Automata consists of States and Transitions. The States represent the Output, i.e the strategy played by the player. The transitions represent the Input, i.e the strategy played by the opponent. C C C D D D Example: Tit for Tat

Outline of Argument Limit the number of states an automaton is aloud to have. Present two complex strategies, in which any deviation is retaliated by defecting forever. Show that playing the strategies utilizes all of the states, thus no states are left to count the rounds. Question: What should be the bounds on the number of states?

Trivial Lower Bound Lemma: An automaton with less then n states, cannot count to n. Corollary: If both of the automata are bounded to have less then n states, then ‘Tit for Tat’ is an equilibrium. If one of the automata has less then n states, then ‘Tit for Tat’ is an equilibrium when the ‘smart’ player defects in the last round.

Upper Bound Theorem 1: If both size bounds are at least 2 n-1, then, the only equilibrium is the one in which both players defect in all the rounds. Proof: Bottom up dynamic programming on the decision tree of each player.

Main Result If at least one of the state bounds in the n- round prisoner’s dilemma is bounded by 2 O(εn), then, for large enough n, there is a (mixed) equilibrium with average payoff for each player at least 3-ε.

Equilibrium – General Description The strategies are mixed. The support of each, consists of 2 d pure strategies, represented by a string {C,D} d. d<<n. First the players will exchange their strings. The players will collaborate, while periodically make sure that the opponent remembers their string. A deviation would be retaliated by defecting for ever.

The Equilibrium Automaton d announce Cooperation + Checking All states contain also a retaliating edge that leads to an ever defecting state.

Technical Points I Problem: Some ‘business cards’ are more advantageous than others. Therefore it is better to have a ‘business card’ with many ‘D’. Solution: After the first segment, comes a ‘fix up’ segment: For i=1…d rounds both players play C iff the i’th letter in both their cards is C. Now in the first 2d steps, the average payoff is 1.75, for every business card.

d Announce Cooperation + Checking Fix up d The Equilibrium Automaton

Technical Points II Problem: The Checking segment might not be fair, making one business card more advantageous then another. Solution: The Checking segment is done by xoring the C’s and the D’s. Thus on average the payments are the same for all business cards.

Technical Points III Problem: If the opponent is playing according to his equilibrium strategy, then he will never encounter the punitive edge of a state. A ‘cheating’ automaton might be able to use this in order to save states, and then defect in the last round. Suppose player 1 has a state q in which he plays C and expects C, and another state p in which he plays C and expects D. A cheating player can unite both states and save a state.

The ‘Honest’ Player’s Automaton C C D Punitive State. p C q D DC The ‘Cheating’ Player’s Automaton C C p D

The Solution After the business card exchange, both players play in unison, i.e. each player, when plays C expects C and when plays D expects D. Therefore the problematic scenario does not occur.

Generalization Let G be an arbitrary game and let p = (p 1,p 2 ) be a point in the individually rational region realized by pure strategies. For every ε>0, there is c>0, N>0 such that for all n>N, in the n round repeated game G played by automata. If at least one of the automata is bounded by 2 cn then there exists an equilibrium with average payoff of at least p-ε.