Presentation is loading. Please wait.

Presentation is loading. Please wait.

CS 188: Artificial Intelligence Fall 2009 Lecture 10: MDPs 9/29/2009 Dan Klein – UC Berkeley Many slides over the course adapted from either Stuart Russell.

Similar presentations


Presentation on theme: "CS 188: Artificial Intelligence Fall 2009 Lecture 10: MDPs 9/29/2009 Dan Klein – UC Berkeley Many slides over the course adapted from either Stuart Russell."— Presentation transcript:

1 CS 188: Artificial Intelligence Fall 2009 Lecture 10: MDPs 9/29/2009 Dan Klein – UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1

2 Announcements  P2: Due Wednesday  P3: MDPs and Reinforcement Learning is up!  W2: Out late this week 2

3 Recap: MDPs  Markov decision processes:  States S  Actions A  Transitions P(s’|s,a) (or T(s,a,s’))  Rewards R(s,a,s’) (and discount  )  Start state s 0  Quantities:  Policy = map of states to actions  Episode = one run of an MDP  Utility = sum of discounted rewards  Values = expected future utility from a state  Q-Values = expected future utility from a q-state a s s, a s,a,s’ s’ 3 [DEMO – MDP Quantities]

4 Recap: Optimal Utilities  The utility of a state s: V * (s) = expected utility starting in s and acting optimally  The utility of a q-state (s,a): Q * (s,a) = expected utility starting in s, taking action a and thereafter acting optimally  The optimal policy:  * (s) = optimal action from state s 4 a s s’ s, a (s,a,s’) is a transition s,a,s’ s is a state (s, a) is a q-state

5 Recap: Bellman Equations  Definition of utility leads to a simple one-step lookahead relationship amongst optimal utility values: Total optimal rewards = maximize over choice of (first action plus optimal future)  Formally: a s s, a s,a,s’ s’ 5

6 Practice: Computing Actions  Which action should we chose from state s:  Given optimal values V?  Given optimal q-values Q?  Lesson: actions are easier to select from Q’s! 6 [DEMO – MDP action selection]

7 Value Estimates  Calculate estimates V k * (s)  Not the optimal value of s!  The optimal value considering only next k time steps (k rewards)  As k  , it approaches the optimal value  Almost solution: recursion (i.e. expectimax)  Correct solution: dynamic programming 7 [DEMO -- V k ]

8 Memoized Recursion?  Recurrences (basically truncated expectimax):  Cache all function call results so you never repeat work  What happened to the evaluation function? 8

9 Value Iteration  Problems with the recursive computation:  Have to keep all the V k * (s) around all the time  Don’t know which depth  k (s) to ask for when planning  Solution: value iteration  Calculate values for all states, bottom-up  Keep increasing k until convergence 9

10 Value Iteration  Idea:  Start with V 0 * (s) = 0, which we know is right (why?)  Given V i *, calculate the values for all states for depth i+1:  Throw out old vector V i *  Repeat until convergence  This is called a value update or Bellman update  Theorem: will converge to unique optimal values  Basic idea: approximations get refined towards optimal values  Policy may converge long before values do 10

11 Convergence*  Define the max-norm:  Theorem: For any two approximations U and V  I.e. any distinct approximations must get closer to each other, so, in particular, any approximation must get closer to the true U and value iteration converges to a unique, stable, optimal solution  Theorem:  I.e. once the change in our approximation is small, it must also be close to correct 11

12 Utilities for a Fixed Policy  Another basic operation: compute the utility of a state s under a fixed (generally non-optimal) policy  Define the utility of a state s, under a fixed policy  : V  (s) = expected total discounted rewards (return) starting in s and following   Recursive relation (one-step look- ahead / Bellman equation):  (s) s s,  (s) s,  (s),s’ s’ 12 [DEMO – Right-Only Policy]

13 Policy Evaluation  How do we calculate the V’s for a fixed policy?  Idea one: turn recursive equations into updates  Idea two: it’s just a linear system, solve with Matlab (or whatever) 13

14 Policy Iteration  Alternative approach:  Step 1: Policy evaluation: calculate utilities for some fixed policy (not optimal utilities!) until convergence  Step 2: Policy improvement: update policy using one- step look-ahead with resulting converged (but not optimal!) utilities as future values  Repeat steps until policy converges  This is policy iteration  It’s still optimal!  Can converge faster under some conditions 14

15 Policy Iteration  Policy evaluation: with fixed current policy , find values with simplified Bellman updates:  Iterate until values converge  Policy improvement: with fixed utilities, find the best action according to one-step look-ahead 15

16 Comparison  Both compute same thing (optimal values for all states)  In value iteration:  Every pass (or “backup”) updates both utilities (explicitly, based on current utilities) and policy (implicitly, based on current utilities)  Tracking the policy isn’t necessary; we take the max  In policy iteration:  Several passes to update utilities with fixed policy  After policy is evaluated, a new policy is chosen  Together, these are dynamic programming for MDPs 16

17 Asynchronous Value Iteration*  In value iteration, we update every state in each iteration  Actually, any sequences of Bellman updates will converge if every state is visited infinitely often  In fact, we can update the policy as seldom or often as we like, and we will still converge  Idea: Update states whose value we expect to change: If is large then update predecessors of s

18 Reinforcement Learning  Reinforcement learning:  Still have an MDP:  A set of states s  S  A set of actions (per state) A  A model T(s,a,s’)  A reward function R(s,a,s’)  Still looking for a policy  (s)  New twist: don’t know T or R  I.e. don’t know which states are good or what the actions do  Must actually try actions and states out to learn [DEMO] 18

19 Example: Animal Learning  RL studied experimentally for more than 60 years in psychology  Rewards: food, pain, hunger, drugs, etc.  Mechanisms and sophistication debated  Example: foraging  Bees learn near-optimal foraging plan in field of artificial flowers with controlled nectar supplies  Bees have a direct neural connection from nectar intake measurement to motor planning area 19

20 Example: Backgammon  Reward only for win / loss in terminal states, zero otherwise  TD-Gammon learns a function approximation to V(s) using a neural network  Combined with depth 3 search, one of the top 3 players in the world  You could imagine training Pacman this way…  … but it’s tricky! (It’s also P3) 20

21 Passive Learning  Simplified task  You don’t know the transitions T(s,a,s’)  You don’t know the rewards R(s,a,s’)  You are given a policy  (s)  Goal: learn the state values (and maybe the model)  I.e., policy evaluation  In this case:  Learner “along for the ride”  No choice about what actions to take  Just execute the policy and learn from experience  We’ll get to the active case soon  This is NOT offline planning! 21

22 Example: Direct Estimation  Episodes: x y (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (3,3) right -1 (4,3) exit +100 (done) (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (4,2) exit -100 (done) V(1,1) ~ (92 + -106) / 2 = -7 V(3,3) ~ (99 + 97 + -102) / 3 = 31.3  = 1, R = -1 +100 -100 22 [DEMO – Optimal Policy]

23 Model-Based Learning  Idea:  Learn the model empirically (rather than values)  Solve the MDP as if the learned model were correct  Better than direct estimation?  Empirical model learning  Simplest case:  Count outcomes for each s,a  Normalize to give estimate of T(s,a,s’)  Discover R(s,a,s’) the first time we experience (s,a,s’)  More complex learners are possible (e.g. if we know that all squares have related action outcomes, e.g. “stationary noise”) 23

24 Example: Model-Based Learning  Episodes: x y T(, right, ) = 1 / 3 T(, right, ) = 2 / 2 +100 -100  = 1 (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (3,3) right -1 (4,3) exit +100 (done) (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (4,2) exit -100 (done) 24

25 Recap: Model-Based Policy Evaluation  Simplified Bellman updates to calculate V for a fixed policy:  New V is expected one-step-look- ahead using current V  Unfortunately, need T and R 25  (s) s s,  (s) s,  (s),s’ s’

26 Sample Avg to Replace Expectation?  Who needs T and R? Approximate the expectation with samples (drawn from T!) 26  (s) s s,  (s) s1’s1’ s2’s2’ s3’s3’

27 Model-Free Learning  Big idea: why bother learning T?  Update V each time we experience a transition  Frequent outcomes will contribute more updates (over time)  Temporal difference learning (TD)  Policy still fixed!  Move values toward value of whatever successor occurs: running average! 27  (s) s s,  (s) s’

28 Example: TD Policy Evaluation Take  = 1,  = 0.5 (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (3,3) right -1 (4,3) exit +100 (done) (1,1) up -1 (1,2) up -1 (1,3) right -1 (2,3) right -1 (3,3) right -1 (3,2) up -1 (4,2) exit -100 (done) 28

29 Problems with TD Value Learning  TD value leaning is model-free for policy evaluation (passive learning)  However, if we want to turn our value estimates into a policy, we’re sunk:  Idea: learn Q-values directly  Makes action selection model-free too! a s s, a s,a,s’ s’ 29


Download ppt "CS 188: Artificial Intelligence Fall 2009 Lecture 10: MDPs 9/29/2009 Dan Klein – UC Berkeley Many slides over the course adapted from either Stuart Russell."

Similar presentations


Ads by Google