Download presentation

Presentation is loading. Please wait.

Published byDenzel Hacking Modified over 2 years ago

1
Cooperation via Policy Search and Unconstrained Minimization Brendan and Yifang Feb 24 2015

2
Paper: Learning to Cooperate via Policy Search Peshkin, Leonid and Kim, Kee-Eung and Meuleau, Nicolas and Kaelbling, Leslie Pack. 2000.

3
Introduction Previously we’ve been concerned with single agents, oblivious environments and one-to-one value functions Impractical for large/diverse environments Impractical for complicated/reactive systems More generalizable approach desirable RL Chapter 8 introduced function approximation to build more accurate value functions with small samples Peshkin et al. introduce multiple agents and world state information asymmetry Wide range of potential new applications Particularly relevant to game theory Compatible with Connectionism models

4
Game Theory “Game theory is the study of the ways in which interacting choices of economic agents produce outcomes with respect to the preferences (or utilities) of those agents, where the outcomes in question might have been intended by none of the agents.” - Stanford Encyclopedia of Philosophy Example – Volunteer’s Dilemma: N agents share a common reward if at least one agent volunteers, but volunteering has an associated cost (e.g. power goes out and at least one person has to make an expensive sat-phone call to electric company to get power restored) Cooperative Multi-agent Imperfect world state information Analytic solutions

5
Multi-Agent Environments

6
Paper Specifics Paper adopts Partially Observable Identical Payoff Stochastic Game model with some caveats Picked to be interesting Allows theoretic guarantees Requires non-trivial distributed algorithm Some game theory applications

7
Identical Payoff Stochastic Game (IPSG)

8
Reward Questions

9
Reward Questions Cont. Question: “ Its title is about cooperation. Does the reward function r reflect the idea of cooperation? ” – Jiyun Luo Paper’s definition - “Cooperative games are those in which both agents share the same payoff structure.” Identical payoff stochastic games meet this requirement by definition Cooperation effect seemingly evident in some example – e.g. soccer More generally – “a cooperative game is a game where groups of players ("coalitions") may enforce cooperative behavior, hence the game is a competition between coalitions of players, rather than between individual players.” Typically entails communication and enforcement not present in model Similarly, joint strategies are not explicit in this model

10
Partially Observable IPSG (POIPSG)

11
Asymmetric Information Question Question: “The paper investigates the problem in which "agents all receive the shared reward signal, but have incomplete, unreliable, and generally different perceptions of the world state." How important is this problem? What are the real world applications to this model? ” – Yuankai Economics: Equal shared holders in a company all benefit from its success, but may have different world state information Politics: Citizens all have a common goal, but differ in knowledge Shared rewards are common (ie volunteer’s dilemma), but identical rewards are somewhat unrealistic Model arguably lacks communication and enforcement necessary for more complex economic problems; potentially representable with observation function?

12
Model Assumptions State space is discrete Agents have finite memory Important for realism Paper outlines a finite state controller for each agent, but uses fully reactive policies in practice Factored controller (ie distributed opposed to centralized), each agent has its own sub-policy Not compatible with simple communication models Learning is simultaneous Actions happen simultaneously coordinated by the reward result Policy defines probability of action as a continuous differential function

13
Joint Controller vs. Factored Controller

14
Control Question “What is the major difference (mathematically) between central control vs. distributed control of factored actions?” – Dr. Yang In general, see previous In the case of gradient descent, none Action probabilities of each agent (and consequently the distributed partial derivatives) are independent. As such, weight updates can be distributed without cooperation

15
Algorithm Question

16
Simple Artificial Neuron

17
Algorithm - REINFORCE

18
Algorithm – Gradient Descent for Policy Search

19
Gradient Descent for Policy Search Cont.

20
Algorithm – Distributed Gradient Descent Now that we can calculate the gradient for each trial, using gradient descent is trivial Gradient descent implementations will be covered later To distribute the action determination to individual agents (consistent with our model), simply have each agent perform gradient descent locally Each agent given some set of initial weights Each agent adjusts their own weight given their local observation and the reward for each round

21
Guarantees For factored controllers, distributed gradient descent is equivalent to joint gradient descent Thus, Distributed Gradient Descent finds a local optimum Centralized gradient descent finds a local optimum simply because it’s a descent algorithm that stops at an optimum Distributed is equivalent to joint Every strict Nash equilibrium is a local optimum for gradient descent Equivalence isn’t two-way (discussed later)

22
Nash Equilibrium

23
NE Question “ If we are modeling two agents in a system, say a user and a machine. To reach Nash equilibrium, can we restrict what the user can or should do in the process? Is this reasonable? If not, that means that the user can do whatever he/she wants, and acts not optimal, any comments on the Nash equilibrium under such condition? How can the machine still optimize for both parties? ” A NE is two-sided by definition. If a user can improve their return (ie isn’t acting optimally) it isn’t a NE. Rational agents are typically a good assumption, though it often requires a very sophisticated model to reflect real life Partial observability may be a good way model mostly rational agents ie a user interacting with a search engine The machine may have a dominant strategy irrespective of the user actions, but it’s generally not interesting to model such problems with multiple agents

24
NE Equivalence Question “Can you explain when a local optima for gradient descent would not be a Nash equilibrium?” – Brad Gradient descent is only guaranteed to search locally One agent may have a better strategy (far removed) that improves theirs payoff regardless of other agents “We can construct a value function V (w1, w2) such that for some c, V (·, c) has two modes, one at V (a, c) and the other at V (b, c), such that V (b, c)> V (a, c). Further assume that V (a, ·) and V (b, ·) each have global maxima V (a, c) and V (b, c). Then V (a, c) is a local optimum that is not a Nash equilibrium”

25
Example - Toy Coordination problem s1 s2 s3 s4 s5 s6 +10 -10 +5 ;

26
Experimental Results

27
Example - Soccer 6x5 grid with 2 randomly places agents and 1 defender Agents can {North, South, East, West, Stay, Pass} Agents observe who possesses ball and status of surrounding Defenders can be Random, Greedy (goes out to block ball), or Defensive(stays in goal to block)

28
Soccer optimal solution question “ The paper shows a distributed learning algorithm in cooperative multi-agent domain by a 2-agents soccer game example. It points out that, an algorithm which can achieve optimal payoff for all agents may not be possible in general. Why not? Can you explain this in the class? ” – Sicong Solving POIPSG completely is intractable Analytical solutions are sometimes possible

29
Q-learning

30
Experimental Results – Defensive Opponents

31
Experimental Results – Greedy Opponents

32
Experimental Results – Random Opponents

33
Experimental Results – Mixed Opponents

34
Tractability Question “The paper gives the example of tractable soccer game with two learning agents playing against one opponent with fixed strategy and then it mentions that it becomes difficult to even store the Q table. The question is, what can be the possible to solve complex environments that are useful for practical applications, if any? Discussion on this?” – Tavish Recall: Q-learning was mainly introduced as motivation Removing the Q table is favor of a compact function approximation based on sampling is much more practical for large examples and often not too bad.

35
Conclusions Reinforcement learning can easily generalized to include multiple agents and applied to game theory Distributed gradient descent has nice guarantees and performs well in POIPSG Distributed gradient descent can produce “cooperation” against consistent and coordinated opponents

36
Convex Optimization: Section 9.1 Stephen Boyd and Lieven Vandenberghe. 2004.

37
Problem Our goal is to solve the unconstrained minimization problem: Where f(x) is convex, and twice continuously differentiable. In this case, a necessary and sufficient condition for x to be optimal is So what is convex?

38
Convex, strictly convex f(x) is called convex, if f(x) is called strictly convex, if

39
Strongly convex f(x) is twice continuously differentiable, and it is strongly convex, if it satisfy the following condition where is Hessian matrix

40
Strong convex, cont’d Thus, the m should be smaller than the smallest Eigen value of the Hessian matrix.

41
Lower bound of According to Taylor theorem, we will have Then we have It means when is small, the f(x) is near to its optimal.

42
Upper-bound of Similar to lower bound, has an upper bound.

43
Condition number of sublevel sets We define the width of a context set C in the direction q, q is a unit vector. Further, we have minimum width and maximum width The condition number of a context set C is

44
Question 1: What is Hessian matrix?(From Brad) Answer: It is second-order derivative matrix Question 2: Given convexity, does local optimal guarantee to be global optimal? Answer: Yes, local optimal guarantees global optimal, if the objective function is continuously differentiable.

45
Question 3: Convexity is a strong assumption. Does the real- world problem always satisfy this assumption? What if it does not stand? Answer: The real-world problems do not always satisfy this assumption. We adopt some other strategies: follow the other direction other than gradient-descent direction, with a small probability. Question 4: Can you explain the math symbol in Section 9.1?(From Jiyun)

46
Convex Optimization: Sections 9.2 and 9.3 Stephen Boyd and Lieven Vandenberghe. 2004.

47
Descent Methods

48
General descent method

49
Descent Direction Question

50
Exact Line Search

51
Backtracking Line Search

52
Backtracking Stopping Condition Question

53
Backtracking Line Search Ex.

54
Backtracking Stopping Condition Question “ Could you show us graphically what it looks like when the stopping condition of the bactracking line search holds?” - Brad

55
Backtracking Stop Ex.

56
Backtracking Parameter Choice Question

57
Line Search Question

58
Gradient Descent

59
Line Search Questions Cont. “Since the setting is convex and twice continuously differentiable, can you compare the exact line search, backtracking line search, and the Newton’s Method?” – Sicong Newton’s method is a descent method, but it relies on the Hessian rather than the gradient so it’s using a different search vector Convergence guarantees rely on an additional self-concordance assumption “When would one use exact, vs backtracking line search?” - Brad Exact line search typically reduces the required number of steps (not guaranteed) However, computing the exact minimum often not feasible or efficient “What are the major differences between gradient method with exact line search and gradient method with backtracking line search? Please illustrate with an example how they differ in reaching the point of convergence.” - Tavish

60
Newton’s Method

61
Gradient Descent with Exact Line Search

62
Gradient Descent with Backtracking Line Search

63
Exact vs. Backtracking Speed

64
Conclusions Based on toy experimental results Gradient method offers roughly linear convergence Choice of backtracking parameters has a noticeable but not dramatic effect Gradient descent is advantageous due to its simplicity Gradient descent may converge slowly in certain cases

Similar presentations

OK

CS 478 - Bayesian Learning1 Bayesian Learning. CS 478 - Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.

CS 478 - Bayesian Learning1 Bayesian Learning. CS 478 - Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on indian water purifier industry Ppt on next generation 2-stroke engine diagram Ppt on new technology in computer science Ppt on basics of object oriented programming Ppt on elections in india downloads Ppt on remote controlled screw jack Ppt on revolution of the earth and seasons kids Ppt on area of a circle Lc esi ms ppt online Ppt on ecology and environment