6/2/2001 Cooperative Agent Systems: Artificial Agents Play the Ultimatum Game Steven O. Kimbrough Presented at FMEC 2001, Oslo Joint work with Fang Zhong.

Slides:



Advertisements
Similar presentations
An Introduction to Game Theory Part V: Extensive Games with Perfect Information Bernhard Nebel.
Advertisements

Ultimatum Game Two players bargain (anonymously) to divide a fixed amount between them. P1 (proposer) offers a division of the “pie” P2 (responder) decides.
GAME THEORY.
1 On the Methodology of Inequity Aversion Theory.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
1 Chapter 14 – Game Theory 14.1 Nash Equilibrium 14.2 Repeated Prisoners’ Dilemma 14.3 Sequential-Move Games and Strategic Moves.
1 Game Theory. By the end of this section, you should be able to…. ► In a simultaneous game played only once, find and define:  the Nash equilibrium.
Game Theory and Computer Networks: a useful combination? Christos Samaras, COMNET Group, DUTH.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
Game Theory: Inside Oligopoly
Game Theory Game theory is an attempt to model the way decisions are made in competitive situations. It has obvious applications in economics. But it.
Automata-based adaptive behavior for economic modeling using game theory Rawan Ghnemat, Khalaf Khatatneh, Saleh Oqeili Al-Balqa’ Applied University, Al-Salt,
The Evolution of Cooperation within the Iterated Prisoner’s dilemma on a Social Network.
17. (A very brief) Introduction to game theory Varian, Chapters 28, 29.
Maynard Smith Revisited: Spatial Mobility and Limited Resources Shaping Population Dynamics and Evolutionary Stable Strategies Pedro Ribeiro de Andrade.
On Quantum Walks and Iterated Quantum Games G. Abal, R. Donangelo, H. Fort Universidad de la República, Montevideo, Uruguay UFRJ, RJ, Brazil.
The Evolution of Conventions H. Peyton Young Presented by Na Li and Cory Pender.
An Introduction to Game Theory Part I: Strategic Games
Satisfaction Equilibrium Stéphane Ross. Canadian AI / 21 Problem In real life multiagent systems :  Agents generally do not know the preferences.
1 Duke PhD Summer Camp August 2007 Outline  Motivation  Mutual Consistency: CH Model  Noisy Best-Response: QRE Model  Instant Convergence: EWA Learning.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Chapter 12 Choices Involving Strategy McGraw-Hill/Irwin Copyright © 2008 by The McGraw-Hill Companies, Inc. All Rights Reserved.
Game-Theoretic Approaches to Multi-Agent Systems Bernhard Nebel.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Review Midterm3/21 3/7.
A Memetic Framework for Describing and Simulating Spatial Prisoner’s Dilemma with Coalition Formation Sneak Review by Udara Weerakoon.
CS522: Algorithmic and Economic Aspects of the Internet Instructors: Nicole Immorlica Mohammad Mahdian
Selfish Caching in Distributed Systems: A Game-Theoretic Analysis By Byung-Gon Chun et al. UC Berkeley PODC’04.
Artificial Agents Play the Beer Game Eliminate the Bullwhip Effect and Whip the MBAs Steven O. Kimbrough D.-J. Wu Fang Zhong FMEC, Philadelphia, June 2000;
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Nash Q-Learning for General-Sum Stochastic Games Hu & Wellman March 6 th, 2006 CS286r Presented by Ilan Lobel.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Review Midterm3/23 3/2.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
1 Lecture 2: Negotiating Strategy Professor Keith Chen.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Review Midterm3/19 3/5.
1 On the Agenda(s) of Research on Multi-Agent Learning by Yoav Shoham and Rob Powers and Trond Grenager Learning against opponents with bounded memory.
The Agencies Method for Coalition Formation in Experimental Games John Nash (University of Princeton) Rosemarie Nagel (Universitat Pompeu Fabra, ICREA,
Complexity and the Nascent Revolution in Economics Lancaster University Dec 9, 2009 W. Brian Arthur External Professor, Santa Fe Institute.
Artificial Intelligence in Game Design Problems and Goals.
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
A Game-Theoretic Approach to Strategic Behavior. Chapter Outline ©2015 McGraw-Hill Education. All Rights Reserved. 2 The Prisoner’s Dilemma: An Introduction.
ISMA 2004 Workshop on Internet Signal Processing (WISP) 1 Perspectives on Resource Allocation Kameswari Chebrolu, Bhaskaran Raman, Ramesh R. Rao November.
Learning in Multiagent systems
1 Dr. Itamar Arel College of Engineering Electrical Engineering & Computer Science Department The University of Tennessee Fall 2009 August 24, 2009 ECE-517:
Plan for Today: Neoliberal Institutionalism & Concluding Liberalism 1. Complete group activity reporting. 2. Survey neoliberal solutions to the Prisoner’s.
Monetary Economics Game and Monetary Policymaking.
Presented by Qian Zou.  The purpose of conducting the experiments.  The methodology for the experiments.  The Experimental Design : Cohesion Experiments.
Classifying Attributes with Game- theoretic Rough Sets Nouman Azam and JingTao Yao Department of Computer Science University of Regina CANADA S4S 0A2
Intermediate Microeconomics
Transit price negotiation: repeated game approach Sogea 23 Mai 2007 Nancy, France D.Barth, J.Cohen, L.Echabbi and C.Hamlaoui
Bargaining Whoever offers to another a bargain of any kind, proposes to do this. Give me that which I want, and you shall have this which you want …; and.
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Linking multi-agent simulation to experiments in economy Re-implementing John Duffy’s model of speculative learning agents.
Session 2a, 10th June 2008 ICT-MobileSummit 2008 Copyright E3 project, BUPT Autonomic Joint Session Admission Control using Reinforcement Learning.
The Evolution of Fairness PSC 120 Jeff Schank. Fairness People engage in fair exchanges of resources even when it would benefit them more to act unfairly.
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
Bargaining as Constraint Satisfaction Simple Bargaining Game Edward Tsang
Artificial intelligence
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Evolving Reactive NPCs for the Real-Time Simulation Game.
Equilibrium transitions in stochastic evolutionary games Dresden, ECCS’07 Jacek Miękisz Institute of Applied Mathematics University of Warsaw.
Lecture 1 on Bargaining Setting the Agenda This lecture focuses on the well known problem of how to split the gains from trade or, more generally, mutual.
Intermediate Microeconomics Game Theory and Oligopoly.
Joint Power and Channel Minimization in Topology Control: A Cognitive Network Approach J ORGE M ORI A LEXANDER Y AKOBOVICH M ICHAEL S AHAI L EV F AYNSHTEYN.
Bargaining games Econ 414. General bargaining games A common application of repeated games is to examine situations of two or more parties bargaining.
Lec 23 Chapter 28 Game Theory.
The highly intelligent virtual agents for modeling financial markets G. Yang 1, Y. Chen 2 and J. P. Huang 1 1 Department of Physics, Fudan University.
Unit 4 SOCIAL INTERACTIONS.
Choices Involving Strategy
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Presentation transcript:

6/2/2001 Cooperative Agent Systems: Artificial Agents Play the Ultimatum Game Steven O. Kimbrough Presented at FMEC 2001, Oslo Joint work with Fang Zhong and D.J. Wu

6/2/2001 Research Motivation How design and control cooperative agent systems in strategic situation. How well do different identify-centric agents perform against each other. How well do various adaptive mechanism perform. Value of intelligence: What intelligence buys you?

6/2/2001 Methodology Adaptive artificial agents play iterated ultimatum game. Ultimatum game is the most fundamental building block for negotiation (e.g., Croson, 1996) Reinforcement learning (a simple version) Regimes of play –Two agents play against each other –Populations of different type of agents

6/2/2001 One-shot Ultimatum Game Two players A and B. Player A has endowment of N. Player A offers x  [0, N] (N = 100 in this study) Player B can either accept the offer or reject the offer. A x B (N-x, x) (0,0) Accept Reject

6/2/2001 One-shot Ultimatum Game (Cont.) Classical Game Theory –Player A offer a tiny amount , and player B will always accept this offer. –Infinite number of Nash Equilibria along the line of x + y = N. Behavior Game Theory –Human beings in the lab do not behave as classical game theory predicted (e.g, people tends to be fair, and reject offers that do not meet their threshold amounts of share).

6/2/2001 Repeated Ultimatum Game A “supergame” consists of iterations of the ultimatum game. Indefinite episodes –Agents do not know how many iterations are yet to come. No single best strategy for the repeated ultimatum game.

6/2/2001 Reinforcement Learning Favoring actions producing better results. Estimating the values of state-action pairs. Sample-average for estimation/evaluation.  -greedy for selection.

6/2/2001 Reinforcement Learning (Cont.) Algorithm Initialize Q(s, a) = 0 Repeat for each episode Choose action a from current state Receive immediate payoff r, and arrive at the next state. Q(s, a) <- Q B (s, a)*(k-1)/k + r/k Until n episodes have been played.

6/2/2001 Experiment 0: Repeated One- Shot Game Agents have no memory of past actions. Agents find the game-theoretic result. No cooperation among agents.

6/2/2001 Experiment 1: Learning Agent Against Fixed Rules Fixing player B’s strategy IF (currentOffer >= p * Endowment) Accept currentOffer. ELSE Reject currentOffer. 0 < p < 1

6/2/2001 Experiment 1 (Cont.) Player A will propose an offer no greater than his last offer if player B accepted his last offer. Player A eventually learns the value of p, and proposes only the amount of pN.

6/2/2001 Experiment 2: Learning Agent Against Dynamic Rules The value of p is changing along the game playing period. Agent A can track the change very well given enough time periods The values of p in different episodes p

6/2/2001 Experiment 3: Learning Agent Against Rotating Rules The value of p is changing with a rotating pattern, i.e. p t-1 =.40, p t =.50, p t+1 =.60. Player A converges to a proposal of 60 which the highest value of p * 100. Memory of at least one previous move might lead player A track the rotated rules.

6/2/2001 Experiment 4: Learning Simultaneously Both agents have memory of one previous move. Player B chooses the value of p for each episode according to: IF b t –1 is “accept” THEN p t = d t-1 / N ELSE p t  [0, N] / N

6/2/2001 Experiment 4 (Cont.) Decision-making process using finite automata Agent A: Reject Accept Reject dd*ad*a d* b Accept

6/2/2001 Experiment 4 (Cont.) Agent B: D D C D p = d p* C C p

6/2/2001 Experiment 4 - Result Cooperation emerges through co-evolution within 2000 episodes. Player A converges at proposing 55 or 56, and correspondingly, player B converges at setting his lower limit at 55 or 56.

6/2/2001 Value of Intelligence Will smart agents be able to do better than dumb ones through learning? Experiment: –5a: A population of smart agents play against a population of various dumb agents –5b: A population of smart agents play against each other and against a population of various dumb agents.

6/2/2001 Experiment 5a: One Smart Agent vs. Multiple Dumb Agents Three types of dumb agents using fixed rules: –db1: demand/accept 70 or higher; –db2: demand/accept 50 or higher; –db3: demand/accept 30 or higher. Smart agent learns via reinforcement learning. There is 25 percent possibility that a smart agent can be chosen to play the game. Tracking the changing population of dumb agents for each generation.

6/2/2001 Experiment 5a : Process Draw one smart agent with 25 percent possibility; otherwise draw one dumb agent randomly in proportional to their frequency. Draw another dumb agent randomly in proportional to their frequency. Decide the role of each agent (proposer or responder). Agents play the one-shot game against each other. Go to the first step until a certain number of games, e.g episodes, has been completed. Update frequency of the dumb agent.

6/2/2001 Experiment 5a – Results. Fair dumb (db2: demand/accept 50 or higher) agents take over the dumb agent population. Smart agents learn to be fair.

6/2/2001 Experiment 5a – Result (Cont.)

6/2/2001 Experiment 5a – Result (Cont.)

6/2/2001 Experiment 5b: Multiple Smart Agents vs. Dumb Agents Smart agents can play against each other.

6/2/2001 Experiment 5b (Cont.)

6/2/2001 Comparison of 5a & 5b

6/2/2001 Impact of Memory Repeat experiment 5a and 5b, but introduce different memory size for each experiment.

6/2/2001 Conclusions Artificial agents using reinforcement learning are able to play the ultimatum game efficiently and effectively. Agent intelligence and memory have impacts on performance. Agent-based approach replicates and explains real human behavior better.

6/2/2001 Future Research Toward cooperative agent systems in strategic situations in virtual communities, especially in electronic commerce such as in supply chains. Currently investigating two versions of the trust games: “The classical economic trust game” vs. “The Mad Mex Game”. Comments?