Learning in Multiagent systems

Slides:



Advertisements
Similar presentations
News and Notes 4/13 HW 3 due now HW 4 distributed today, due Thu 4/22 Final exam is Mon May 3 11 AM Levine 101 Today: –intro to evolutionary game theory.
Advertisements

Ultimatum Game Two players bargain (anonymously) to divide a fixed amount between them. P1 (proposer) offers a division of the “pie” P2 (responder) decides.
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
ECON 100 Tutorial: Week 9 office: LUMS C85.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Dispute Settlement Mechanism The role of dispute settlement mechanism –information gathering and dispatching, not enforcement of trade arrangements Main.
Congestion Games with Player- Specific Payoff Functions Igal Milchtaich, Department of Mathematics, The Hebrew University of Jerusalem, 1993 Presentation.
Joint Strategy Fictitious Play Sherwin Doroudi. “Adapted” from J. R. Marden, G. Arslan, J. S. Shamma, “Joint strategy fictitious play with inertia for.
1 Chapter 14 – Game Theory 14.1 Nash Equilibrium 14.2 Repeated Prisoners’ Dilemma 14.3 Sequential-Move Games and Strategic Moves.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
An Introduction to... Evolutionary Game Theory
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
Learning in games Vincent Conitzer
What is a game?. Game: a contest between players with rules to determine a winner. Strategy: a long term plan of action designed to achieve a particular.
Game Theoretic Aspect in Human Computation Presenter: Chien-Ju Ho
The Evolution of Cooperation within the Iterated Prisoner’s dilemma on a Social Network.
Game Theory Lecture 8.
Maynard Smith Revisited: Spatial Mobility and Limited Resources Shaping Population Dynamics and Evolutionary Stable Strategies Pedro Ribeiro de Andrade.
ECO290E: Game Theory Lecture 4 Applications in Industrial Organization.
The Evolution of Conventions H. Peyton Young Presented by Na Li and Cory Pender.
An Introduction to Game Theory Part I: Strategic Games
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Algoritmi per Sistemi Distribuiti Strategici
Temporal Action-Graph Games: A New Representation for Dynamic Games Albert Xin Jiang University of British Columbia Kevin Leyton-Brown University of British.
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
Algorithmic Game Theory Nicola Gatti and Marcello Restelli {ngatti, DEI, Politecnico di Milano, Piazza Leonardo da Vinci 32, Milano,
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
6/2/2001 Cooperative Agent Systems: Artificial Agents Play the Ultimatum Game Steven O. Kimbrough Presented at FMEC 2001, Oslo Joint work with Fang Zhong.
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
1 On the Agenda(s) of Research on Multi-Agent Learning by Yoav Shoham and Rob Powers and Trond Grenager Learning against opponents with bounded memory.
CPS Learning in games Vincent Conitzer
5. Alternative Approaches. Strategic Bahavior in Business and Econ 1. Introduction 2. Individual Decision Making 3. Basic Topics in Game Theory 4. The.
NOBEL WP Szept Stockholm Game Theory in Inter-domain Routing LÓJA Krisztina - SZIGETI János - CINKLER Tibor BME TMIT Budapest,
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
Game Theory is Evolving MIT , Fall Our Topics in the Course  Classical Topics Choice under uncertainty Cooperative games  Values  2-player.
1 Economics & Evolution Number 3. 2 The replicator dynamics (in general)
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Game Theory in Wireless and Communication Networks: Theory, Models, and Applications Lecture 4 Evolutional Game Zhu Han, Dusit Niyato, Walid Saad, Tamer.
Topic 3 Games in Extensive Form 1. A. Perfect Information Games in Extensive Form. 1 RaiseFold Raise (0,0) (-1,1) Raise (1,-1) (-1,1)(2,-2) 2.
Equilibrium transitions in stochastic evolutionary games Dresden, ECCS’07 Jacek Miękisz Institute of Applied Mathematics University of Warsaw.
Normal Form Games, Normal Form Games, Rationality and Iterated Rationality and Iterated Deletion of Dominated Strategies Deletion of Dominated Strategies.
Vincent Conitzer CPS Learning in games Vincent Conitzer
Replicator Dynamics. Nash makes sense (arguably) if… -Uber-rational -Calculating.
Evolution of Cooperation in Mobile Ad Hoc Networks Jeff Hudack (working with some Italian guy)
Application of Game Theory to Biology A long-standing puzzle in the study of animal behavior is the prevalence of conventional fights in the competition.
Q 2.1 Nash Equilibrium Ben
Game Theory and Cooperation
Replicator Dynamics.
Communication Complexity as a Lower Bound for Learning in Games
Vincent Conitzer CPS Repeated games Vincent Conitzer
Multiagent Systems Game Theory © Manfred Huber 2018.
Multiagent Systems Repeated Games © Manfred Huber 2018.
Vincent Conitzer Repeated games Vincent Conitzer
Evolving cooperation in one-time interactions with strangers
Vincent Conitzer Learning in games Vincent Conitzer
Vincent Conitzer CPS Learning in games Vincent Conitzer
Phase transitions to cooperation in the prisoner‘s dilemma
Vincent Conitzer CPS Repeated games Vincent Conitzer
Presentation transcript:

Learning in Multiagent systems Prepared by: Jarosław Szymczak Based on: „Fundamentals of Multiagent Systems with NetLogo Examples” by José M Vidal

Scenarios of learning cooperative learning – e.g. each agent has own map, and together with other agents aggregate a global view competitive learning – e.g. each selfish agent tries to maximize own utility by learning about behaviors and weaknesses of other agents agents learn because: they don’t know everything about the environment they don’t know how the other agents behave

The Machine Learning Problem The goal of machine learning research is the development of algorithms that increase the ability of an agent to match a set of inputs to their corresponding outputs (Mitchell, 1997) Input here could be e.g. a set of photos depicting people and output would be set consisting of {man, woman}, the machine learning algorithm will have to learn a proper recognition of photos

The Machine Learning Problem Input set is usually divided into training and testing set, they can be interleaved Graphical representation of MLP:

The Machine Learning Problem Induction bias – some learning algorithms appear to perform better than others in certain domains (e.g. two algorithms can learn perfectly classify + and -, but still have different functions) No free lunch theorem - averages over all possible learning problems there is no learning algorithm that outperforms all others In multiagent scenario some of the fundamental assumptions of machine learning are violated, there is no longer fixed input, it keeps changing because other agents are also learning

Cooperative learning We have given a two robots able to communicate, they can share their knowledge (their capabilities, knowledge about terrain etc.) Sharing the information is really easy if robots are identical, if not we need to somehow model their capabilities to decide which information would be useful for another robot Most systems that share learned knowledge among agents, such as (Stone, 2000), simply assume that all agents have the same capabilities.

Repeated games Nash equilibrium – I choose what is best for me when you are doing what you are doing and you choose what is best for you when I am doing what I am doing In repeated games we have two players facing each other, like e.g. prisoner's dilemma Nash equilibrium is based on assumption of perfectly rational players, in learning in games the assumption is that agents use some kind of algorithm, the theory determines the equilibrium strategy that will be arrived at by the various learning mechanisms and maps these equilibrium to the standard solution concepts, if possible.

Fictitious play Agent remembers everything the other agents have done. E.g.:

Fictitious play Let’s have a look at some theorems: (Nash Equilibrium is Attractor to Fictitious Play). If s is a strict Nash equilibrium and it is played at time t then it will be played at all times greater than t (Fudenberg and Kreps, 1990). (Fictitious Play Converges to Nash). If fictitious play converges to a pure strategy then that strategy must be a Nash equilibrium (Fudenberg and Kreps, 1990).

Fictitious play Infinite cycles problem – we can avoid it by using randomness, here is the example of infinite cycle in fictitious play:

Replicator dynamics This model assumes that the fraction of agents playing a particular strategy replicator dynamics will grow in proportion to how well that strategy performs in the population. A homogeneous population of agents is assumed. The agents are randomly paired in order to play a symmetric game (same strategies and payoffs). It is inspired by biological evolution. Let φt(s) be a number of agents using strategy s at time t, ut(s) be an expected utility for an agent playing strategy s at time t and u(s,s’) be utility that agant playing s receives against agent playing s’. We can define:

Replicator dynamics Let’s have a look at some theorems: (Nash equilibrium is a Steady State). Every Nash equilibrium is a steady state for the replicator dynamics (Fudenberg and Levine, 1998). (Stable Steady State is a Nash Equilibrium). A stable steady state of the replicator dynamics is a Nash equilibrium. A stable steady state is one that, after suffering from a small perturbation, is pushed back to the same steady state by the system’s dynamics (Fudenberg and Levine, 1998). (Asymptotically Stable is Trembling-Hand Nash). An asymptotically stable steady state corresponds to a Nash equilibrium that is trembling-hand perfect and isolated. That is, the stable steady states are a refinement on Nash equilibria - only a few Nash equilibria are stable steady states (Bomze, 1986).

Evolutionary stable strategy An ESS is an equilibrium strategy that can overcome the presence of a small number of invaders. That is, if the equilibrium strategy profile is ω and small number ε of invaders start playing ω’ then ESS states that the existing population should get a higher payoff against the new mixture (εω‘+(1−ε)ω) than the invaders. (ESS is Steady State of Replicator Dynamics). ESS is an asymptotically stable steady state of the replicator dynamics. However, the converse need not be true—a stable state in the replicator dynamics does not need to be an ESS (Taylor and Jonker, 1978).

Replicator dynamics

AWESOME algorithm The abbreviation stands for: Adapt When Every is Stationary, Otherwise Move to Equilibrium

COMING SOON  (THIS AUTUMN) Stochastic games COMING SOON  (THIS AUTUMN)