IJCAI’07 Emergence of Norms through Social Learning Partha Mukherjee, Sandip Sen and Stéphane Airiau Mathematical and Computer Sciences Department University.

Slides:



Advertisements
Similar presentations
1 Reinforcement Learning (RL). 2 Introduction The concept of reinforcement learning incorporates an agent that solves the problem in hand by interacting.
Advertisements

Agent-based Modeling: A Brief Introduction Louis J. Gross The Institute for Environmental Modeling Departments of Ecology and Evolutionary Biology and.
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Genetic Algorithms (Evolutionary Computing) Genetic Algorithms are used to try to “evolve” the solution to a problem Generate prototype solutions called.
DARWIN: Distributed and Adaptive Reputation Mechanism for Wireless Ad- hoc Networks CHEN Xiao Wei, Cheung Siu Ming CSE, CUHK May 15, 2008 This talk is.
6-1 LECTURE 6: MULTIAGENT INTERACTIONS An Introduction to MultiAgent Systems
Joint Strategy Fictitious Play Sherwin Doroudi. “Adapted” from J. R. Marden, G. Arslan, J. S. Shamma, “Joint strategy fictitious play with inertia for.
Game Theory and Computer Networks: a useful combination? Christos Samaras, COMNET Group, DUTH.
Game Theory: Inside Oligopoly
Game Theory. Games Oligopolist Play ▫Each oligopolist realizes both that its profit depends on what its competitor does and that its competitor’s profit.
Learning in games Vincent Conitzer
Stephen McCray and David Courard-Hauri, Environmental Science and Policy Program, Drake University Introduction References 1.Doran, P. T. & Zimmerman,
Class Project Due at end of finals week Essentially anything you want, so long as it’s AI related and I approve Any programming language you want In pairs.
Reinforcement Learning
1. Algorithms for Inverse Reinforcement Learning 2
Fair Division of Indivisible Goods Thomas Kalinowski (Newcastle) Nina Naroditskaya, Toby Walsh (NICTA, UNSW) Lirong Xia (Harvard)
A Computational Characterization of Multiagent Games with Fallacious Rewards Ariel D. Procaccia and Jeffrey S. Rosenschein.
The Evolution of Conventions H. Peyton Young Presented by Na Li and Cory Pender.
Satisfaction Equilibrium Stéphane Ross. Canadian AI / 21 Problem In real life multiagent systems :  Agents generally do not know the preferences.
1 Best-Reply Mechanisms Noam Nisan, Michael Schapira and Aviv Zohar.
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
0 Network Effects in Coordination Games Satellite symposium “Dynamics of Networks and Behavior” Vincent Buskens Jeroen Weesie ICS / Utrecht University.
Centinel tournament ● A deck: the numbers in random order ● A game lasts until no numbers are left in deck ● A game is played like this (first player.
6/2/2001 Cooperative Agent Systems: Artificial Agents Play the Ultimatum Game Steven O. Kimbrough Presented at FMEC 2001, Oslo Joint work with Fang Zhong.
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
Outline MDP (brief) –Background –Learning MDP Q learning Game theory (brief) –Background Markov games (2-player) –Background –Learning Markov games Littman’s.
Nash Q-Learning for General-Sum Stochastic Games Hu & Wellman March 6 th, 2006 CS286r Presented by Ilan Lobel.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
Extending Implicit Negotiation to Repeated Grid Games Robin Carnow Computer Science Department Rutgers University.
Reinforcement Learning Game playing: So far, we have told the agent the value of a given board position. How can agent learn which positions are important?
CPS Learning in games Vincent Conitzer
Game Theory.
MAKING COMPLEX DEClSlONS
CISC 235: Topic 6 Game Trees.
History-Dependent Graphical Multiagent Models Quang Duong Michael P. Wellman Satinder Singh Computer Science and Engineering University of Michigan, USA.
Derivative Action Learning in Games Review of: J. Shamma and G. Arslan, “Dynamic Fictitious Play, Dynamic Gradient Play, and Distributed Convergence to.
Learning in Multiagent systems
Introduction Many decision making problems in real life
Learning BlackJack with ANN (Aritificial Neural Network) Ip Kei Sam ID:
Some Analysis of Coloring Experiments and Intro to Competitive Contagion Assignment Prof. Michael Kearns Networked Life NETS 112 Fall 2014.
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
1 Near-Optimal Play in a Social Learning Game Ryan Carr, Eric Raboin, Austin Parker, and Dana Nau Department of Computer Science, University of Maryland.
Linking multi-agent simulation to experiments in economy Re-implementing John Duffy’s model of speculative learning agents.
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
Cilk Pousse James Process CS534. Overview Introduction to Pousse Searching Evaluation Function Move Ordering Conclusion.
Bargaining as Constraint Satisfaction Simple Bargaining Game Edward Tsang
Reinforcement Learning Yishay Mansour Tel-Aviv University.
Institute of Physics Wroclaw University of Technology 28/09/2005 How can statistical mechanics contribute to social sciences? Piotr Magnuszewski, Andrzej.
Decision Making Under Uncertainty CMSC 471 – Spring 2041 Class #25– Tuesday, April 29 R&N, material from Lise Getoor, Jean-Claude Latombe, and.
Vincent Conitzer CPS Learning in games Vincent Conitzer
The Role of Altruistic Punishment in Promoting Cooperation
The Standard Genetic Algorithm Start with a “population” of “individuals” Rank these individuals according to their “fitness” Select pairs of individuals.
Evolving Strategies for the Prisoner’s Dilemma Jennifer Golbeck University of Maryland, College Park Department of Computer Science July 23, 2002.
Computation, The Missing Ingredient in Classical Economics Edward Tsang Centre for Computational Finance and Economic Agents (CCFEA) University of Essex.
Social Norm, Costly Punishment and the Evolution to Cooperation : Theory, Experiment and Simulation Tongkui Yu 1, 2, Shu-Heng Chen 2, Honggang Li 1* 1.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
CS 5751 Machine Learning Chapter 13 Reinforcement Learning1 Reinforcement Learning Control learning Control polices that choose optimal actions Q learning.
Coordination with Linear Equations
Reinforcement Learning in POMDPs Without Resets
Communication Complexity as a Lower Bound for Learning in Games
Convergence, Targeted Optimality, and Safety in Multiagent Learning
Announcements Homework 3 due today (grace period through Friday)
Multiagent Systems Game Theory © Manfred Huber 2018.
CASE − Cognitive Agents for Social Environments
Area Models A strategy for Multiplication
Introduction to RePast and Tutorial I
Mutual support in agent networks
A Deep Reinforcement Learning Approach to Traffic Management
Presentation transcript:

IJCAI’07 Emergence of Norms through Social Learning Partha Mukherjee, Sandip Sen and Stéphane Airiau Mathematical and Computer Sciences Department University of Tulsa, Oklahoma, USA

ALAg-07 Introduction Norm: “a convention as an equilibrium that everyone expects in interactions that have more than one equilibrium” [Young, 1996] Use a population of learning agents to simulate a population that faces a problem modeled by a game and study the emergence of norms

ALAg-07 Example of a norm: picking the side of the road Agents need to decide on one of several equally desirable alternatives. This game can be extended to m actions 4 4 L R RL

ALAg-07 Previous Work Previous work on learning norms assume observation of other interactions between agents.  How norms will emerge if all interactions were private? Social Learning (IJCAI-07): agents play a bimatrix game, at each interaction, an agent plays against another agent, taken at random, in the population  Empirical study: Study effect of population size, number of actions available, effect of learning algorithms, presence of non-learning agents, multiple relatively isolated populations

ALAg-07 Social Learning Population of N learning agents A 2-player, k-action game M M is common knowledge Each agent has a learning algorithm (fixed, intrinsic) to play M as a row or a column player Repeatedly, agents play the game M against an unknown, random opponent.

ALAg-07 Protocol of play For each iteration, for each agent Pick randomly one agent in its neighborhood For each pair, one agent is randomly considered row, the other column player Each agent pick an action, and can observe only the action of the other agent constituting the pair Each agent gets the reward accordingly, and updates its learning mechanism

ALAg-07 Interactions are limited to neighboring agents

ALAg-07 Effect of neighboring size

ALAg-07 Learning Dynamics D=1 D=15 It 145It 355It 480  Driving on the left  Driving on the right

ALAg-07 Influence of non-learners Non-learners use identical strategies D=5

ALAg-07 Influence of non-learners Using different strategies  Driving on the left  Driving on the right D=1 D=15 It 45It 535 It 905

ALAg-07 Conclusion Bottom up process for the emergence of social norms Depends only on private expertise Agents can learn and sustain useful social norms Agent population with smaller neighborhoods converge faster to a norm