1 Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A. Franz J. Kurfess CPE/CSC 580: Intelligent Agents 1.

Slides:



Advertisements
Similar presentations
Concepts of Game Theory I
Advertisements

Concepts of Game Theory II. 2 The prisioners reasoning… Put yourself in the place of prisoner i (or j)… Reason as follows: –Suppose I cooperate… If j.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
SIF8072 Distributed Artificial Intelligence and Intelligent Agents
Multiagent Interactions
Game Theory “Доверяй, Но Проверяй” - Russian Proverb (Trust, but Verify) - Ronald Reagan Mike Shor Lecture 6.
Non-Cooperative Game Theory To define a game, you need to know three things: –The set of players –The strategy sets of the players (i.e., the actions they.
6-1 LECTURE 6: MULTIAGENT INTERACTIONS An Introduction to MultiAgent Systems
1 Game Theory. By the end of this section, you should be able to…. ► In a simultaneous game played only once, find and define:  the Nash equilibrium.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
An Introduction to... Evolutionary Game Theory
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
Infinitely Repeated Games Econ 171. Finitely Repeated Game Take any game play it, then play it again, for a specified number of times. The game that is.
 1. Introduction to game theory and its solutions.  2. Relate Cryptography with game theory problem by introducing an example.  3. Open questions and.
Network Theory and Dynamic Systems Game Theory: Mixed Strategies
Coye Cheshire & Andrew Fiore March 21, 2012 // Computer-Mediated Communication Collective Action and CMC: Game Theory Approaches and Applications.
Multi-player, non-zero-sum games
An Introduction to Game Theory Part I: Strategic Games
Game Theory Part 5: Nash’s Theorem.
2008/02/06Lecture 21 ECO290E: Game Theory Lecture 2 Static Games and Nash Equilibrium.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
Nash Equilibrium: Theory. Strategic or Simultaneous-move Games Definition: A simultaneous-move game consists of: A set of players For each player, a set.
5/16/20151 Game Theory Game theory was developed by John Von Neumann and Oscar Morgenstern in Economists! One of the fundamental principles of.
Chapter 11 What is Utility? A way of representing preferences Utility is not money (but it is a useful analogy) Typical relationship between utility &
GAME THEORY By Ben Cutting & Rohit Venkat. Game Theory: General Definition  Mathematical decision making tool  Used to analyze a competitive situation.
Sep. 5, 2013 Lirong Xia Introduction to Game Theory.
Slide 1 of 13 So... What’s Game Theory? Game theory refers to a branch of applied math that deals with the strategic interactions between various ‘agents’,
Arguments for Recovering Cooperation Conclusions that some have drawn from analysis of prisoner’s dilemma: – the game theory notion of rational action.
A Memetic Framework for Describing and Simulating Spatial Prisoner’s Dilemma with Coalition Formation Sneak Review by Udara Weerakoon.
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
Static Games of Complete Information: Equilibrium Concepts
1 Game Theory Here we study a method for thinking about oligopoly situations. As we consider some terminology, we will see the simultaneous move, one shot.
APEC 8205: Applied Game Theory Fall 2007
Game Theory Here we study a method for thinking about oligopoly situations. As we consider some terminology, we will see the simultaneous move, one shot.
Design of Multi-Agent Systems Teacher Bart Verheij Student assistants Albert Hankel Elske van der Vaart Web site
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
QR 38, 2/22/07 Strategic form: dominant strategies I.Strategic form II.Finding Nash equilibria III.Strategic form games in IR.
Today: Some classic games in game theory
Social Choice Session 7 Carmen Pasca and John Hey.
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
A Game-Theoretic Approach to Strategic Behavior. Chapter Outline ©2015 McGraw-Hill Education. All Rights Reserved. 2 The Prisoner’s Dilemma: An Introduction.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
THE “CLASSIC” 2 x 2 SIMULTANEOUS CHOICE GAMES Topic #4.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
LECTURE 6: MULTIAGENT INTERACTIONS
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
The Science of Networks 6.1 Today’s topics Game Theory Normal-form games Dominating strategies Nash equilibria Acknowledgements Vincent Conitzer, Michael.
Chapter 10: Games and Strategic Behavior
Lecture 5 Introduction to Game theory. What is game theory? Game theory studies situations where players have strategic interactions; the payoff that.
Section 2 – Ec1818 Jeremy Barofsky
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Mohsen Afsharchi Multiagent Interaction. What are Multiagent Systems?
Game theory (Sections )
Arguments for Recovering Cooperation Conclusions that some have drawn from analysis of prisoner’s dilemma: – the game theory notion of rational action.
Advanced Subjects in GT Outline of the tutorials Static Games of Complete Information Introduction to games Normal-form (strategic-form) representation.
Game Theory By Ben Cutting & Rohit Venkat.
Game theory Chapter 28 and 29
LECTURE 6: MULTIAGENT INTERACTIONS
Chapter 15: Game Theory: The Mathematics Lesson Plan of Competition
LECTURE 6: MULTIAGENT INTERACTIONS
Introduction to Game Theory
Game theory Chapter 28 and 29
Decision Theory and Game Theory
Multiagent Systems Game Theory © Manfred Huber 2018.
LECTURE 6: MULTIAGENT INTERACTIONS
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Chapter 15: Game Theory: The Mathematics Lesson Plan of Competition
Presentation transcript:

1 Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A. Franz J. Kurfess CPE/CSC 580: Intelligent Agents 1

2 © Franz J. Kurfess Usage of the Slides ◆ these slides are intended for the students of my CPE/CSC 580 “Intelligent Agents” class at Cal Poly SLO ◆ if you want to use them outside of my class, please let me know ◆ some of them are based on other sources, which are identified and cited ◆ I usually select a subset for each quarter, either by hiding some slides, or creating a “Custom Show” (in PowerPoint) ◆ to view these, go to “Slide Show => Custom Shows”, select the respective quarter, and click on “Show” ◆ To print them, I suggest to use the “Handout” option ◆ 4, 6, or 9 per page works fine ◆ Black & White should be fine; there are few diagrams where color is important

3 © Franz J. Kurfess Course Overview ❖ Introduction  Intelligent Agent, Multi-Agent Systems  Agent Examples ❖ Agent Architectures  Agent Hierarchy, Agent Design Principles ❖ Reasoning Agents  Knowledge, Reasoning, Planning ❖ Learning Agents  Observation, Analysis, Performance Improvement ❖ Multi-Agent Interactions  Agent Encounters, Resource Sharing, Agreements ❖ Communication  Speech Acts, Agent Communication Languages ❖ Collaboration  Distributed Problem Solving, Task and Result Sharing ❖ Agent Applications  Information Gathering, Workflow, Human Interaction, E-Commerce, Embodied Agents, Virtual Environments ❖ Conclusions and Outlook

4 © Franz J. Kurfess Overview Multi-Agent Encounters ❖ Motivation ❖ Objectives ❖ Multi-Agent Systems  co-existence, interaction, action, relationships ❖ Agent Encounters  shared environment, actions, rational action  solution concepts ❖ Resource Sharing ❖ Agreements  self-interest, mutual benefits, negotiation, argumentation ❖ Important Concepts and Terms ❖ Chapter Summary

5 © Franz J. Kurfess Bridge-In

6 © Franz J. Kurfess Pre-Test

7 © Franz J. Kurfess Motivation

8 © Franz J. Kurfess Objectives

9 © Franz J. Kurfess Evaluation Criteria

10 Multi-Agent Basics Multi-agent systems Utilities and preferences

11 [Woolridge 2009] Multi-agent Systems?

12 [Woolridge 2009] Multi-Agent Systems a multi-agent system contains a number of agents that  share an environment possibly also resources  act in an environment  interact through communication  have different “spheres of influence” which may coincide  are linked by other (organizational) relationships

13 [Woolridge 2009] Utilities and Preferences Agents are assumed to be self-interested:  they have preferences over how the environment is Agents have preferences over outcomes  Ω = {ω1, ω2, …} is the set of “outcomes” Preferences are captured by utility functions Utility functions lead to preference orderings over outcomes:

14 Agent Encounters Environment Encounter Models Payoff Matrices Solution Concepts Strategies Examples: Games

15 [Woolridge 2009] Multi-agent Encounters environment model  agents simultaneously choose an action as a result of the actions, an outcome in Ω will result  the actual outcome depends on the combination of actions assumptions  only two agents: Ag = {i, j}  each agent can perform two possible actions C (“cooperate”) D (“defect”)

16 [Woolridge 2009] State Transformer Function environment behavior is given by a state transformer function:

17 [Woolridge 2009] 6-7 Simple Multi-Agent Encounters environment sensitive to actions of both agents: neither agent has any influence in this environment: environment is controlled by j:

18 [Woolridge 2009] 6-7 Simple Multi-Agent Encounters environment sensitive to actions of both agents: neither agent has any influence in this environment: environment is controlled by j:

19 [Woolridge 2009] Rational Action assume that both agents can influence the outcome  utility functions as follows: re-written for our defect/cooperate scenario agent i’s preferences are: “C” is the rational choice for agent i  i prefers all outcomes that arise through C over all outcomes that arise through D

20 [Woolridge 2009] 6-9 Payoff Matrices payoff matrix for the previous scenario in a:  agent i is the column player  agent j is the row player

21 [Woolridge 2009] Solution Concepts How will a rational agent will behave in any given scenario? Answered in solution concepts:  dominant strategy  Nash equilibrium strategy  Pareto optimal strategies  strategies that maximize social welfare 6-10

22 [Woolridge 2009] 6-11 Dominant Strategies We will say that a strategy s i is dominant for player i if no matter what strategy s j agent j chooses, i will do at least as well playing s i as it would doing anything else Unfortunately, there isn’t always a unique undominated strategy

23 [Woolridge 2009] 6-12 (Pure Strategy) Nash Equilibrium In general, we will say that two strategies s 1 and s 2 are in Nash equilibrium if: 1. under the assumption that agent i plays s 1, agent j can do no better than play s 2 ; and  under the assumption that agent j plays s 2, agent i can do no better than play s 1. Neither agent has any incentive to deviate from a Nash equilibrium Unfortunately:  Not every interaction scenario has a Nash equilibrium  Some interaction scenarios have more than one Nash equilibrium

24 [Woolridge 2009] Matching Pennies Players i and j simultaneously choose the face of a coin, either “heads” or “tails” If they show the same face, then i wins, while if they show different faces, then j wins The Matching Pennies Payoff Matrix: 6-13

25 [Woolridge 2009] Mixed Strategies for Matching Pennies NO pair of strategies forms a pure strategy Nash Equilibrium (NE): whatever pair of strategies is chosen, somebody will wish they had done something else The solution is to allow mixed strategies:  play “heads” with probability 0.5  play “tails” with probability 0.5 This is a NE strategy 6-14

26 [Woolridge 2009] Mixed Strategies A mixed strategy has the form  play α 1 with probability p 1  play α 2 with probability p 2  …  play α k with probability p k  such that p 1 + p 2 + ··· + p k =1 6-15

27 [Woolridge 2009] Nash’s Theorem Nash proved that every finite game has a Nash equilibrium in mixed strategies.  unlike the case for pure strategies  this result overcomes the lack of solutions  but there still may be more than one Nash equilibrium 6-16

28 [Woolridge 2009] Pareto Optimality an outcome is said to be Pareto optimal (or Pareto efficient)  if there is no other outcome that makes one agent better off without making another agent worse off if an outcome is Pareto optimal,  then at least one agent will be reluctant to move away from it because this agent will be worse off 6-17

29 [Woolridge 2009] Pareto Optimality and Reasonable Agents if an outcome ω is not Pareto optimal  then there is another outcome ω’ that makes everyone as happy, if not happier, than ω reasonable” agents would agree to move to ω’ in this case  even if I don’t directly benefit from ω’, you can benefit without me suffering

30 [Woolridge 2009] Social Welfare sum of the utilities that each agent gets from outcome ω:  “total amount of money in the system” may be appropriate when the whole system (all agents) has a single owner  the overall benefit of the system is important, not benefits of individuals

31 [Woolridge 2009] Competitive and Zero-Sum Interactions strictly competitive scenarios  preferences of agents are diametrically opposed zero-sum encounters  utilities ad up to zero

32 [Woolridge 2009] Dilemma of Zero Sum Interactions zero sum encounters are bad news  for me to get positive utility you have to get negative utility  the best outcome for me is the worst for you zero sum encounters in real life are very rare  but people tend to act in many scenarios as if they were zero sum

33 [Woolridge 2009] 6-22 The Prisoner’s Dilemma Two men are collectively charged with a crime and held in separate cells, with no way of meeting or communicating. They are told that:  if one confesses and the other does not, the confessor will be freed, and the other will be jailed for three years  if both confess, then each will be jailed for two years Both prisoners know that if neither confesses, then they will each be jailed for one year

34 [Woolridge 2009] 6-23 The Prisoner’s Dilemma Payoff matrix for prisoner’s dilemma: Top left: If both defect, then both get punishment for mutual defection Top right: If i cooperates and j defects, i gets sucker’s payoff of 1, while j gets 4 Bottom left: If j cooperates and i defects, j gets sucker’s payoff of 1, while i gets 4 Bottom right: Reward for mutual cooperation

35 [Woolridge 2009] 6-24 What Should You Do? The individual rational action is defect This guarantees a payoff of no worse than 2, whereas cooperating guarantees a payoff of at most 1 So defection is the best response to all possible strategies: both agents defect, and get payoff = 2 But intuition says this is not the best outcome: Surely they should both cooperate and each get payoff of 3!

36 [Woolridge 2009] Solution Concepts D is a dominant strategy (D, D) is the only Nash equilibrium All outcomes except (D, D) are Pareto optimal (C, C) maximizes social welfare 6-25

37 [Woolridge 2009] 6-26 The Prisoner’s Dilemma for Agents This apparent paradox is the fundamental problem of multi-agent interactions.  It appears to imply that cooperation will not occur in societies of self-interested agents. Real world examples:  nuclear arms reduction (“why don’t I keep mine... ”)  free rider systems — public transport;  in the UK — television licenses. The prisoner’s dilemma is ubiquitous. Can we recover cooperation?

38 [Woolridge 2009] 6-27 Tempting Conclusions Conclusions that some have drawn from this analysis:  the game theory notion of rational action is wrong!  somehow the dilemma is being formulated wrongly Arguments to recover cooperation:  We are not all Machiavelli!  The other prisoner is my twin!  Program equilibria and mediators  The shadow of the future…

39 [Woolridge 2009] 6-27 Arguments for Recovering Cooperation We are not all Machiavelli!  but some may be The other prisoner is my twin!  she may be my evil twin Program equilibria and mediators The shadow of the future…

40 [Woolridge 2009] Program Equilibria The strategy you really want to play in the prisoner’s dilemma is: I’ll cooperate if he will Program equilibria provide one way of enabling this Each agent submits a program strategy to a mediator which jointly executes the strategies. Crucially, strategies can be conditioned on the strategies of the others 6-28

41 [Woolridge 2009] Program Equilibria Consider the following program: IF HisProgram == ThisProgram THEN DO(C); ELSE DO(D); END-IF  Here == is textual comparison  best response: submit the same program giving an outcome of (C, C)  can’t get the sucker’s payoff through this program 6-29

42 [Woolridge 2009] 6-30 The Iterated Prisoner’s Dilemma One answer: play the game more than once  If you know you will be meeting your opponent again, then the incentive to defect appears to evaporate Cooperation is the rational choice in the infinitely repeated prisoner’s dilemma

43 [Woolridge 2009] 6-31 Backwards Induction Problem suppose you both know that you will play the game exactly n times  on round n - 1, you have an incentive to defect to gain that extra bit of payoff  this makes round n – 2 the last “real”, so you have an incentive to defect there, too backwards induction problem  Playing the prisoner’s dilemma with a fixed, finite, pre-determined, commonly known number of rounds, defection is the best strategy

44 [Woolridge 2009] 6-32 Axelrod’s Tournament iterated prisoner’s dilemma against a range of opponents  What strategy should you choose to maximize your overall payoff? Axelrod (1984) investigated this problem,  computer tournament for programs playing the prisoner’s dilemma

45 [Woolridge 2009] 6-33 Strategies in Axelrod’s Tournament ALLD:  “Always defect” — the hawk strategy; TIT-FOR-TAT: 1. On round u = 0, cooperate 2. On round u > 0, do what your opponent did on round u – 1 TESTER:  On 1st round, defect. If the opponent retaliated, then play TIT-FOR-TAT. Otherwise intersperse cooperation and defection. JOSS:  As TIT-FOR-TAT, except periodically defect

46 [Woolridge 2009] 6-34 Recipes for Success in Axelrod’s Tournament Axelrod suggests the following rules for succeeding in his tournament:  Don’t be envious: Don’t play as if it were zero sum!  Be nice: Start by cooperating, and reciprocate cooperation  Retaliate appropriately: Always punish defection immediately, but use “measured” force — don’t overdo it  Don’t hold grudges: Always reciprocate cooperation immediately

47 [Woolridge 2009] 6-35 Game of Chicken slightly different type of encounter  James Dean in Rebel without a Cause: swerving = coop, driving straight = defect difference to prisoner’s dilemma: Mutual defection is most feared outcome  sucker’s payoff is most feared in prisoner’s dilemma Strategies (c,d) and (d,c) are in Nash equilibrium

48 [Woolridge 2009] Solution Concepts there is no dominant strategy  in our sense strategy pairs (C, D) and (D, C) are Nash equilibriums all outcomes except (D, D) are Pareto optimal all outcomes except (D, D) maximize social welfare 6-36

49 [Woolridge 2009] 6-37 Other Symmetric 2 x 2 Games Given the 4 possible outcomes of (symmetric) cooperate/defect games, there are 24 possible orderings on outcomes  CC > i CD > i DC > i DD Cooperation dominates  DC > i DD > i CC > i CD Deadlock. You will always do best by defecting  DC > i CC > i DD > i CD Prisoner’s dilemma  DC > i CC > i CD > i DD Chicken  CC > i DC > i DD > i CD Stag hunt

50 Resource Sharing

51 Agreements Self-interest Mutual benefits Negotiation Argumentation

52 © Franz J. Kurfess

53 © Franz J. Kurfess Post-Test

54 © Franz J. Kurfess Evaluation ❖ Criteria

55 © Franz J. Kurfess Summary Learning Agents

56 © Franz J. Kurfess Important Concepts and Terms ❖ agent ❖ competition ❖ collaboration ❖ coordination ❖ encounter ❖ environment ❖ equilibrium ❖ multi-agent system ❖ Nash equilibrium ❖ Pareto Optimality ❖ payoff matrix ❖ preference ❖ Prisoner’s Dilemma ❖ rational agent ❖ resource ❖ strategy ❖ utility ❖ zero-sum game

57 © Franz J. Kurfess