1 CSC 384 Lecture Slides (c) 2002-2003, C. Boutilier and P. Poupart CSC384: Lecture 25  Last time Decision trees and decision networks  Today wrap up.

Slides:



Advertisements
Similar presentations
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) Nov, 28, 2012.
Advertisements

Making Simple Decisions
ECON 100 Tutorial: Week 9 office: LUMS C85.
Lauritzen-Spiegelhalter Algorithm
Lecture Roger Sutton 21: Revision 1.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
CPSC 502, Lecture 11Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 11 Oct, 18, 2011.
CPSC 322, Lecture 35Slide 1 Finish VE for Sequential Decisions & Value of Information and Control Computer Science cpsc322, Lecture 35 (Textbook Chpt 9.4)
Decision Theory: Single Stage Decisions Computer Science cpsc322, Lecture 33 (Textbook Chpt 9.2) March, 30, 2009.
Temporal Action-Graph Games: A New Representation for Dynamic Games Albert Xin Jiang University of British Columbia Kevin Leyton-Brown University of British.
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) April, 1, 2009.
Non-Linear Problems General approach. Non-linear Optimization Many objective functions, tend to be non-linear. Design problems for which the objective.
10/24  Exam on 10/26 (Lei Tang and Will Cushing to proctor)
CPSC 322, Lecture 30Slide 1 Reasoning Under Uncertainty: Variable elimination Computer Science cpsc322, Lecture 30 (Textbook Chpt 6.4) March, 23, 2009.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
CPSC 322, Lecture 12Slide 1 CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12 (Textbook Chpt ) January, 29, 2010.
Ensemble Learning: An Introduction
Decision Making Under Uncertainty Russell and Norvig: ch 16, 17 CMSC421 – Fall 2005.
1 Civil Systems Planning Benefit/Cost Analysis Scott Matthews Courses: /
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) April, 12, 2010.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
16.5 Introduction to Cost- based plan selection Amith KC Student Id: 109.
10/22  Homework 3 returned; solutions posted  Homework 4 socket opened  Project 3 assigned  Mid-term on Wednesday  (Optional) Review session Tuesday.
DAST 2005 Week 4 – Some Helpful Material Randomized Quick Sort & Lower bound & General remarks…
Decision Making Under Uncertainty Russell and Norvig: ch 16, 17 CMSC421 – Fall 2003 material from Jean-Claude Latombe, and Daphne Koller.
CMSC 671 Fall 2003 Class #26 – Wednesday, November 26 Russell & Norvig 16.1 – 16.5 Some material borrowed from Jean-Claude Latombe and Daphne Koller by.
CSCI 347 – Data Mining Lecture 01 – Course Overview.
Key Stone Problem… Key Stone Problem… next Set 22 © 2007 Herbert I. Gross.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Reasoning Under Uncertainty: Bayesian networks intro CPSC 322 – Uncertainty 4 Textbook §6.3 – March 23, 2011.
Lecture 22 More NPC problems
Department of Computer Science Undergraduate Events More
Unsupervised Learning: Clustering Some material adapted from slides by Andrew Moore, CMU. Visit for
Monte Carlo Methods So far we have discussed Monte Carlo methods based on a uniform distribution of random numbers on the interval [0,1] p(x) = 1 0  x.
1 Computing Trust in Social Networks Huy Nguyen Lab seminar April 15, 2011.
BUSINESS PERFORMANCE MANAGEMENT
Announcements Project 4: Ghostbusters Homework 7
QM Spring 2002 Business Statistics Probability Distributions.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
Copyright © 2009 Cengage Learning 22.1 Chapter 22 Decision Analysis.
Week 9 Lecture 2 Decision Trees.
CS 8625 High Performance Computing Dr. Hoganson Copyright © 2003, Dr. Ken Hoganson CS8625 Class Will Start Momentarily… CS8625 High Performance.
Inference Algorithms for Bayes Networks
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
1 1 Slide © 2004 Thomson/South-Western Simulation n Simulation is one of the most frequently employed management science techniques. n It is typically.
2.3. Value of Information: Decision Trees and Backward Induction.
Preferences I give robot a planning problem: I want coffee
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
1 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart CSC384: Lecture 24  Last time Variable elimination, decision trees  Today decision.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
1 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart CSC384: Lecture 16  Last time Searching a Graphplan for a plan, and relaxed plan heuristics.
1 Automated Planning and Decision Making 2007 Automated Planning and Decision Making Prof. Ronen Brafman Various Subjects.
Processing a Decision Tree Connecticut Electronics.
Decision Making ECE457 Applied Artificial Intelligence Spring 2007 Lecture #10.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Recap of L09: Normative Decision Theory
Decisions under uncertainty and risk
Nevin L. Zhang Room 3504, phone: ,
Maximum Expected Utility
L09: Car Buyer Example Joe is going to buy a used car, which could be good with probability 0.8 or a lemon with probability 0.2. Joe's profit will be.
ECE457 Applied Artificial Intelligence Fall 2007 Lecture #10
ECE457 Applied Artificial Intelligence Spring 2008 Lecture #10
Hidden Markov Models Part 2: Algorithms
Decision Maths Dijkstra’s Algorithm.
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CAP 5636 – Advanced Artificial Intelligence
Decision Theory: Single Stage Decisions
CS 188: Artificial Intelligence
Presentation transcript:

1 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart CSC384: Lecture 25  Last time Decision trees and decision networks  Today wrap up decision nets and course summary

2 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Sample Decision Network Disease TstResult Chills Fever BloodTstDrug U optional

3 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Policies  A policy is associated with every decision node telling what to do given the value of the parents.  A policy for BT might be: δ BT (c,f) = bt δ BT (c,~f) = ~bt δ BT (~c,f) = bt δ BT (~c,~f) = ~bt Chills Fever BloodTst

4 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Value of a Policy  Value of a policy δ is the expected utility given that decision nodes are set according to δ  If we set all of the chance nodes, the decision nodes will be set by the policy δ and the utility obtained will be set by the utility node.  Every setting of the chance nodes has a certain probability—given by the CPTs in the decision net.  Thus we can compute the expected value of a policy  Value of δ : EU( δ ) = Σ X P(X, δ (X)) U(X, δ (X)) Probability of this setting of the chance nodes Utility obtained

5 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Value of a Policy EU( δ ) = Σ X P(X, δ (X)) U(X, δ (X)) Note the one needs to have the value of the decision nodes to compute the probability of the setting of the chance node Similarly, the utility might depend on the setting of both types of nodes. Disease TstResult BloodTst Drug U

6 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Optimal Policies  An optimal policy is a policy δ * such that EU( δ* ) ≥ EU( δ ) for all policies δ

7 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Computing the Best Policy  We can work backwards  First compute optimal policy for last decision D. for each asst to parents its parents and each setting of D we compute the expected value of choosing that value of D. The policy mapping is always to map each setting of the parents to the value of D that had maximum expected value. Since all decision nodes are parents of D, setting D and its parents yields a BN with a single UTILITY node.

8 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Computing the Best Policy  TRICK: we can include the utility node’s mapping (setting of parents to utility value) in the collection of input factors, then eliminate ALL variables to obtain the EXPECTED VALUE (without having to first compute the Probability then the Expectation.)

9 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Computing the Best Policy Pr(U) = Σ A,B,C P(A,B,C) Util(U,B,C) = Σ A,B,C P(C|B)P(B|A)P(A)Util(U,B,C) Expt(U) = Pr(U) X Val(U) U C B A Expt(U) = Σ A,B,C P(A,B,C) Util(B,C) = Σ A,B,C P(C|B)P(B|A)P(A)Util(B,C) U is a variable with numbers as values. Now just eliminate all variables. Util(B,C) is the value of U given a setting of its parents. Final result is a single value!

10 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Computing the Best Policy  Next compute policy for previous decision C given policy δ D just determined for last decision D. δ D can be treated as being an ordinary CPT with 0/1 probabilities, and D an ordinary BN variable. With δ D treated this way, when we set C and all of its parents we again obtain an ordinary Bayes net with only a single utility node. Now VE can again be used to compute the expected value of every choice of C give a setting of its parents. And a policy δc for C can be produced that maps C to it’s the value that yielded highest expected value for each setting of its parents. Then we continue to decision before C.

11 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Decision Network Notes  Decision networks commonly used by decision analysts to help structure decision problems  Much work put into computationally effective techniques to solve these common trick: replace the decision nodes with random variables at outset and solve a plain Bayes net (a subtle but useful transformation)  Complexity much greater than BN inference we need to perform a number of variable eliminations to compute the policy for each node. one VE problem for each setting of decision node parents and decision node value

12 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart A Detailed Decision Net Example  Setting: you want to buy a used car, but there’s a good chance it is a “lemon” (i.e., prone to breakdown). Before deciding to buy it, you can take it to a mechanic for inspection. S/he will give you a report on the car, labelling it either “good” or “bad”. A good report is positively correlated with the car being sound, while a bad report is positively correlated with the car being a lemon.  The report costs $50 however. So you could risk it, and buy the car without the report.  Owning a sound car is better than having no car, which is better than owning a lemon.

13 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Car Buyer’s Network Lemon Report InspectBuy U l ~l 0.5 g b n l i ~l i l ~i ~l ~i Rep: good,bad,none b l -600 b ~l 1000 ~b l -300 ~b~l -300 Utility -50 if inspect

14 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Evaluate Last Decision: Buy (1)  EU(B|I,R) = Σ L P(L,I,R,B) U(L,B,I) = Σ L P(L)P(R|L,I) U(L,B,I)  I = i, R = g: EU(buy) = P(l)P(g|l,i)U(l,buy,i) + P(~l)P(g|~l,i)U(~l,buy,i) =.5*.2* *.9*950 = EU(~buy) = P(l)P(g|l,i)U(l,~buy,i) +P(~l)P(g|~l,i)U(~l,~buy,i) =.5*.2* *.9*-350 = So optimal δ Buy (i,g) = buy

15 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Evaluate Last Decision: Buy (2)  I = ~i, R = g  I = i, R = b  I = ~i, R = b  I = i, R = n  I = ~i, R = n  Now we have policy for BUY  And can compute policy for Inspect.  I’ll post the details to the web site.

16 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Course Summary—Study Guide.  Main emphasis in exam will be questions that involve showing how particular algorithms operate.  There will also be some questions about concepts.  At most one question involving a proof (you should be able to get a very good mark even if you don’t do this question).  You should know how to do questions of the type that appeared in the two term tests.  Only material covered in lecture and your assignments will be on the test.

17 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Course Summary—Study Guide.  I will post a more detailed study guide on the web.

18 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart Course Summary—Study Guide.  Search: