Simple search methods for finding a Nash equilibrium Ryan Porter, Eugene Nudelman, and Yoav Shoham Games and Economic Behavior, Vol. 63, Issue 2. pp. 642-661,

Slides:



Advertisements
Similar presentations
Simple Search Methods for Finding a Nash Equilibrium
Advertisements

S. Ceppi, N. Gatti, and N. Basilico DEI, Politecnico di Milano Computing Bayes-Nash Equilibria through Support Enumeration Methods in Bayesian Two-Player.
Heuristic Search techniques
Ch 4. Heuristic Search 4.0 Introduction(Heuristic)
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Adopt Algorithm for Distributed Constraint Optimization
Game Theory Assignment For all of these games, P1 chooses between the columns, and P2 chooses between the rows.
1 Constraint Satisfaction Problems A Quick Overview (based on AIMA book slides)
Continuation Methods for Structured Games Ben Blum Christian Shelton Daphne Koller Stanford University.
1 University of Southern California Keep the Adversary Guessing: Agent Security by Policy Randomization Praveen Paruchuri University of Southern California.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Methods of Proof Chapter 7, second half.. Proof methods Proof methods divide into (roughly) two kinds: Application of inference rules: Legitimate (sound)
Chapter 3: The Efficiency of Algorithms Invitation to Computer Science, Java Version, Third Edition.
6.896: Topics in Algorithmic Game Theory Lecture 11 Constantinos Daskalakis.
Game Theoretical Insights in Strategic Patrolling: Model and Analysis Nicola Gatti – DEI, Politecnico di Milano, Piazza Leonardo.
Joint Strategy Fictitious Play Sherwin Doroudi. “Adapted” from J. R. Marden, G. Arslan, J. S. Shamma, “Joint strategy fictitious play with inertia for.
Multiple Criteria Decision Analysis with Game-theoretic Rough Sets Nouman Azam and JingTao Yao Department of Computer Science University of Regina CANADA.
IBM Labs in Haifa © 2005 IBM Corporation Adaptive Application of SAT Solving Techniques Ohad Shacham and Karen Yorav Presented by Sharon Barner.
P.J. Healy California Institute of Technology Learning Dynamics for Mechanism Design An Experimental Comparison of Public Goods Mechanisms.
Algorithms for solving two- player normal form games Tuomas Sandholm Carnegie Mellon University Computer Science Department.
Satisfaction Equilibrium Stéphane Ross. Canadian AI / 21 Problem In real life multiagent systems :  Agents generally do not know the preferences.
Planning under Uncertainty
INFORMS 2006, Pittsburgh, November 8, 2006 © 2006 M. A. Zinkevich, AICML 1 Games, Optimization, and Online Algorithms Martin Zinkevich University of Alberta.
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Tractable Computational Methods for Finding Nash Equilibria of Perfect-Information Position Auctions David Robert Martin Thompson Kevin Leyton-Brown Department.
1 Refining the Basic Constraint Propagation Algorithm Christian Bessière and Jean-Charles Régin Presented by Sricharan Modali.
1 Computing Nash Equilibrium Presenter: Yishay Mansour.
Chapter 3: The Efficiency of Algorithms Invitation to Computer Science, C++ Version, Third Edition Additions by Shannon Steinfadt SP’05.
Chapter 3: The Efficiency of Algorithms
Complexity of Mechanism Design Vincent Conitzer and Tuomas Sandholm Carnegie Mellon University Computer Science Department.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
Computing Equilibria Christos H. Papadimitriou UC Berkeley “christos”
Minimax strategies, Nash equilibria, correlated equilibria Vincent Conitzer
CPS 170: Artificial Intelligence Game Theory Instructor: Vincent Conitzer.
Computing Equilibria Christos H. Papadimitriou UC Berkeley “christos”
MAKING COMPLEX DEClSlONS
Changing Perspective… Common themes throughout past papers Repeated simple games with small number of actions Mostly theoretical papers Known available.
Mechanisms for Making Crowds Truthful Andrew Mao, Sergiy Nesterko.
林偉楷 Taiwan Evolutionary Intelligence Laboratory.
A quantum protocol for sampling correlated equilibria unconditionally and without a mediator Iordanis Kerenidis, LIAFA, Univ Paris 7, and CNRS Shengyu.
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 11.
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Chih-Ming Chen, Student Member, IEEE, Ying-ping Chen, Member, IEEE, Tzu-Ching Shen, and John K. Zao, Senior Member, IEEE Evolutionary Computation (CEC),
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
CP Summer School Modelling for Constraint Programming Barbara Smith 2. Implied Constraints, Optimization, Dominance Rules.
Ásbjörn H Kristbjörnsson1 The complexity of Finding Nash Equilibria Ásbjörn H Kristbjörnsson Algorithms, Logic and Complexity.
Schreiber, Yevgeny. Value-Ordering Heuristics: Search Performance vs. Solution Diversity. In: D. Cohen (Ed.) CP 2010, LNCS 6308, pp Springer-
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Algorithms for solving two-player normal form games
A local search algorithm with repair procedure for the Roadef 2010 challenge Lauri Ahlroth, André Schumacher, Henri Tokola
Arc Consistency CPSC 322 – CSP 3 Textbook § 4.5 February 2, 2011.
Liang, Introduction to Java Programming, Sixth Edition, (c) 2007 Pearson Education, Inc. All rights reserved Chapter 23 Algorithm Efficiency.
M9302 Mathematical Models in Economics Instructor: Georgi Burlakov 0.Game Theory – Brief Introduction Lecture
5.1.Static Games of Incomplete Information
Adaptive Inlining Keith D. CooperTimothy J. Harvey Todd Waterman Department of Computer Science Rice University Houston, TX.
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
R. Brafman and M. Tennenholtz Presented by Daniel Rasmussen.
Data Driven Resource Allocation for Distributed Learning
Extensive-form games and how to solve them
Constraint Propagation
Structured Models for Multi-Agent Interactions
Multiagent Systems Game Theory © Manfred Huber 2018.
Chapter 3: The Efficiency of Algorithms
Constraint Programming and Backtracking Search Algorithms
Chapter 3: The Efficiency of Algorithms
Multiagent Systems Repeated Games © Manfred Huber 2018.
Normal Form (Matrix) Games
Algorithm Course Algorithms Lecture 3 Sorting Algorithm-1
Presentation transcript:

Simple search methods for finding a Nash equilibrium Ryan Porter, Eugene Nudelman, and Yoav Shoham Games and Economic Behavior, Vol. 63, Issue 2. pp , 2004 Georgia Kastidou David R. Cheriton School of Computer Science University of Waterloo

Outline Problem Contribution Algorithm for 2-player game  Experimental Results Algorithm for n-player game  Experimental Results Conclusions Future Work Take Home Message

Problem we will tackle… Consider a n-player normal-form game G=  N, (A i ), (u i )  where:  N=[1,..n] is the set of players  A i =[a i1,…,a imi ] is the set of actions available to player i  u i :A 1 x…xA n  R is the utility function for each player  Each player selects a mixed strategy from the set of available strategies:  “Support” of a mixed strategy p i is the set of all actions a i in A i such that p i (a i )>0.  x=(x 1,..,x n ): x i is the size of agent’s i support  The expected utility for player i for a strategy profile p=(p 1,...,p n ) is: where: Problem: Design an algorithm which can find a Nash Equilibrium for a normal form game  A strategy profile p* in P is a Nash Equilibrium (NE) if:

Development of Algorithms Two extremes:  1 st : Aims to design a low complexity algorithm Gain deep insight in the structure of the problem Design highly specialized algorithms  2 nd : Aims to design more simple algorithms with an “acceptable” complexity Identify shallow heuristics Hoping that given the increasing computing power they will be sufficient. focus on more “common” problems Which one is better?  Although the first is more interesting, in practice in a number of cases the second is preferred either because is simpler or because it outperforms the first. Note: There are cases of optimal algorithms that have never been implemented because they are too complicate.

Background/Related Work Nash equilibrium is an important concept in game theory  “little is known about the problem of computing a sample NE in a normal-form game” Normal form game is guaranteed to have at least one NE It does not fall into a standard complexity class (Papadimitriou, 2001)  it cannot be cast as a decision problem.

Background/Related Work 2-player game:  Lemke-Howson, (1964)  Dickhaut and Kaplan (1991) Enumerates all possible pairs of supports for 2-play game For each pair it solves a feasibility program n-player game:  Simplicial Subdivision (van der Laan et al., 1987) Approximates a fixed point of a function which is defined on a simplotope  Govindan and Wilson (2003) first perturbs a game to one that has a known equilibrium, and then traces the solution back to the original game as the magnitude of the perturbation approaches zero.

Background/Related Work GAMUT: a “recently” (2004) introduced computational testbed for game theory “Run the GAMUT:A Comprehensive Approach to Evaluating Game-Theoretic Algorithms” by E. Nudelman et al.

What the author propose and what’s their contribution? Propose:  heuristic-based algorithms for 2-player games and for n-player games explore the space of support profiles using a backtracking procedure to instantiate the support for each player separately.  test using a variety of different distributions Use of GAMUT (computational testbed for game theory) Contribution:  in a big number of cases the proposed algorithms outperform the algorithm of Lemke-Howson on 2- player games and the Siplicial Subdivision on n-player games.

The Proposed Algorithms explore “support” profiles:  pure strategies played with nonzero probability use backtracking procedures are biased towards simple solutions  preference for small supports  based on the observation that a number of games in the past proved to have at least one simple solution. e.g. for n = 2, the probability that there exists a NE consistent with a particular support profile varies inversely with the size of the supports, and is zero for unbalanced support profiles.

Proposed Algorithm for 2-players game

υ i : expected utility of agent i in an equilibrium

Proposed Algorithm for 2-players game The first two classes of constraints require that:  each player must be indifferent between all actions within his/her support, and  must not strictly prefer an action outside of his/her support.

Experimental results The authors consider games from a number of different distributions D18: most common one D5, D6, and D7 are also important distributions

Experimental Results Algorithms for 2-player games Experiment-Setup:  2-player, 300-action games drawn from 24 of GAMUT’s 2- player distributions.  executed on 100 games drawn from each distribution. First diagram:  compares the unconditional median running times of the algorithms,  might reflect the fact that there is a greater than 50% chance that the distribution will generate a game with a pure Second diagram:  Compares the percentage of instances solved Third diagram:  the average running time conditional on solving an instance

Experimental Results Algorithms for 2-player games Compares the unconditional median running times of the algorithms. (“Might reflect the fact that there is a greater than 50% chance that the distribution will generate a game with a pure”)

Experimental Results Algorithms for 2-player games Compares the percentage of instances solved

Experimental Results Algorithms for 2-player games Compares the average running time conditional on solving an instance (unconditional average running time)

Experimental Result Algorithms for 2-player games Compare the scaling behavior as the number of actions increases (unconditional average running time)

Experimental Result Algorithms for 2-player games Covariance Games  neither of the algorithms solved any of the games in another “Covariance Game” distribution in which ρ =−0.9,

Proposed Algorithm for n-players games Uses a general backtracking algorithm to solve a constraint satisfaction problem (CSP) for each support size profile The variables in each CSP are:  the supports S i, and  the domain of each S i is the set of supports of size x i. Constraints:  no agent plays a conditionally dominated action.

Proposed Algorithm for n-players games IRSDS:  Input a domain for each player’s support.  For each agent whose support has been instantiated the domain contains only that instantiated support,  For each other agent i it contains all supports of size x i that were not eliminated in a previous call to this procedure.  On each pass of the repeat-until loop, every action found in at least one support in a player’s domain is checked for conditional domination. If a domain becomes empty after the removal of a conditionally dominated action,  the current instantiations of the Recursive-Backtracking are inconsistent, and IRSDS returns failure.  IRSDS repeats until it either returns failure or iterates through all actions of all players without finding a dominated action.

Proposed Algorithm for n-players games

IRSDS R-B IRSDS R-B IRSDS R-B IRSDS R-B … … IRSDS R-B R-B: Recursive Backtracking IRSDS: Iterated Removal of Strictly Dominated Strategies Failed IRSDS RBT Algorithm 2 Failed R-B IRSDS R-B For all x=(x1,..xn) sorted in increasing order first by: and then by:

Experimental Results: n-player games Experiment-Setup:  6-player, 5-action games drawn from 22 of GAMUT’s n-player distributions. 15,625 outcomes and 93,750 payoffs  executed on 100 games drawn from each distribution. First diagram:  compares the unconditional median running times of the algorithms,  might reflect the fact that there is a greater than 50% chance that the distribution will generate a game with a pure Second diagram:  Compares the percentage of instances solved Third diagram:  the average running time conditional on solving an instance

Experimental Results: n-player games Compares the unconditional median running times of the algorithms Compares the percentage of instances solved

Experimental Results: n-player games Compares the average running time conditional on solving an instance

Experimental Results: n-player games Compare the scaling behavior: number of players constant at 6 number of actions varies. (unconditional average running time) Compare the scaling behavior: number of players varies, number of actions constant 5. (unconditional average running time)

Experimental Results Percentage of Pure Strategy NE (2-player game) n-player game

Experimentals Results Average measure of support balanced 2-player, 300-action games 6-player, 5-action games

Conclusions Propose algorithms that use backtracking approaches to search the space of support profiles, favoring supports that are small and balanced. Both algorithms outperform the current state of the art. The most difficult games  “Covariance Game” model, as the covariance approaches its minimal value  hard because authors found that: as the covariance decreases, the number of equilibria decreases, and the equilibria that do exist are more likely to have support sizes near one half of the number of actions

Future Work Employ more sophisticated CSP techniques Explore local search, in which the state space is the set of all possible supports, and the available moves are to add or delete an action from the support of a player Study the games that are generated by the Covariance Game distribution

Take home message Studying the results of complicated problems can lead to observations that although might not provide ideas to find optimal solutions can provide insights on how to improve current approaches. The selection of the tests and the parameters that will be examined very important.  Not only because they can show that your algorithm is working…  E.g. “Covariance Game” model might proved a good starting point for game theoretic algorithms