Department of Computer Science Undergraduate Events More https://my.cs.ubc.ca/students/development/eventshttps://my.cs.ubc.ca/students/development/events.

Slides:



Advertisements
Similar presentations
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) Nov, 28, 2012.
Advertisements

CPSC 322, Lecture 4Slide 1 Search: Intro Computer Science cpsc322, Lecture 4 (Textbook Chpt ) Sept, 11, 2013.
CPSC 502, Lecture 11Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 11 Oct, 18, 2011.
Planning under Uncertainty
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
CPSC 322, Lecture 35Slide 1 Finish VE for Sequential Decisions & Value of Information and Control Computer Science cpsc322, Lecture 35 (Textbook Chpt 9.4)
CPSC 322, Lecture 19Slide 1 Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt ) February, 23, 2009.
Decision Theory: Single Stage Decisions Computer Science cpsc322, Lecture 33 (Textbook Chpt 9.2) March, 30, 2009.
CPSC 322, Lecture 37Slide 1 Finish Markov Decision Processes Last Class Computer Science cpsc322, Lecture 37 (Textbook Chpt 9.5) April, 8, 2009.
CPSC 322, Lecture 4Slide 1 Search: Intro Computer Science cpsc322, Lecture 4 (Textbook Chpt ) January, 12, 2009.
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) April, 1, 2009.
CPSC 322, Lecture 30Slide 1 Reasoning Under Uncertainty: Variable elimination Computer Science cpsc322, Lecture 30 (Textbook Chpt 6.4) March, 23, 2009.
CPSC 322, Lecture 11Slide 1 Constraint Satisfaction Problems (CSPs) Introduction Computer Science cpsc322, Lecture 11 (Textbook Chpt 4.0 – 4.2) January,
CPSC 322, Lecture 2Slide 1 Representational Dimensions Computer Science cpsc322, Lecture 2 (Textbook Chpt1) January, 6, 2010.
CPSC 322, Lecture 23Slide 1 Logic: TD as search, Datalog (variables) Computer Science cpsc322, Lecture 23 (Textbook Chpt 5.2 & some basic concepts from.
CPSC 322, Lecture 12Slide 1 CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12 (Textbook Chpt ) January, 29, 2010.
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) April, 12, 2010.
CPSC 322, Lecture 17Slide 1 Planning: Representation and Forward Search Computer Science cpsc322, Lecture 17 (Textbook Chpt 8.1 (Skip )- 8.2) February,
CPSC 322, Lecture 17Slide 1 Planning: Representation and Forward Search Computer Science cpsc322, Lecture 17 (Textbook Chpt 8.1 (Skip )- 8.2) February,
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
CPSC 322, Lecture 22Slide 1 Logic: Domain Modeling /Proofs + Top-Down Proofs Computer Science cpsc322, Lecture 22 (Textbook Chpt 5.2) March, 8, 2010.
CPSC 322, Lecture 32Slide 1 Probability and Time: Hidden Markov Models (HMMs) Computer Science cpsc322, Lecture 32 (Textbook Chpt 6.5) March, 27, 2009.
Department of Computer Science Undergraduate Events More
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
Department of Computer Science Undergraduate Events More
CPSC 322, Lecture 35Slide 1 Value of Information and Control Computer Science cpsc322, Lecture 35 (Textbook Chpt 9.4) April, 14, 2010.
CPSC 322, Lecture 24Slide 1 Reasoning under Uncertainty: Intro to Probability Computer Science cpsc322, Lecture 24 (Textbook Chpt 6.1, 6.1.1) March, 15,
CPSC 502, Lecture 13Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 13 Oct, 25, 2011 Slide credit POMDP: C. Conati.
Slide 1 Logic: Domain Modeling /Proofs + Top-Down Proofs Jim Little UBC CS 322 – CSP October 22, 2014.
Slide 1 Constraint Satisfaction Problems (CSPs) Introduction Jim Little UBC CS 322 – CSP 1 September 27, 2014 Textbook §
Computer Science CPSC 322 Lecture 4 Search: Intro (textbook Ch: ) 1.
CPSC 322, Lecture 22Slide 1 Logic: Domain Modeling /Proofs + Top-Down Proofs Computer Science cpsc322, Lecture 22 (Textbook Chpt 5.2) Oct, 26, 2010.
Department of Computer Science Undergraduate Events More
Computer Science CPSC 502 Lecture 14 Markov Decision Processes (Ch. 9, up to 9.5.3)
CPSC 322, Lecture 32Slide 1 Probability and Time: Hidden Markov Models (HMMs) Computer Science cpsc322, Lecture 32 (Textbook Chpt 6.5.2) Nov, 25, 2013.
Department of Computer Science Undergraduate Events More
Decision Theoretic Planning. Decisions Under Uncertainty  Some areas of AI (e.g., planning) focus on decision making in domains where the environment.
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
CPSC 322, Lecture 4Slide 1 Search: Intro Computer Science cpsc322, Lecture 4 (Textbook Chpt ) Sept, 12, 2012.
Department of Computer Science Undergraduate Events More
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) Oct, 11, 2013.
Department of Computer Science Undergraduate Events More
Decision Theory: Single & Sequential Decisions. VE for Decision Networks. CPSC 322 – Decision Theory 2 Textbook §9.2 April 1, 2011.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) Nov, 13, 2013.
Markov Decision Processes AIMA: 17.1, 17.2 (excluding ), 17.3.
CPSC 322, Lecture 22Slide 1 Logic: Domain Modeling /Proofs + Top-Down Proofs Computer Science cpsc322, Lecture 22 (Textbook Chpt 5.2) Oct, 30, 2013.
Please Also Complete Teaching Evaluations
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 2
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Planning Under Uncertainty Computer Science cpsc322, Lecture 11
Decision Theory: Single Stage Decisions
Probability and Time: Markov Models
UBC Department of Computer Science Undergraduate Events More
Probability and Time: Markov Models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 2
Decision Theory: Sequential Decisions
Decision Theory: Single Stage Decisions
Probability and Time: Markov Models
Please Also Complete Teaching Evaluations
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Logic: Domain Modeling /Proofs + Computer Science cpsc322, Lecture 22
Probability and Time: Markov Models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Department of Computer Science Undergraduate Events More Simba Technologies Tech Talk: Big Data: Volume, Variety & Velocity Wed., Jan 7 5:30 pm DMP 110 Deloitte Career Talk: What is IT Consulting Wed., Jan 14 5:30 pm DMP 110 CS Speed Mentoring & Townhall Event Thurs., Jan 15 5:30 pm Rm X860, ICICS/CS Bldg. CPSC 422, Lecture 2Slide 1

CPSC 422, Lecture 2Slide 2 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 2 Jan, 7, 2015

CPSC 422, Lecture 23 Lecture Overview Value of Information and Value of Control Markov Decision Processes (MDPs) Formal Specification and example Define a policy for an MDP

CPSC 422, Lecture 2Slide 4 Cpsc 322 Big Picture Environment Problem Query Planning Deterministic Stochastic Search Arc Consistency Search Var. Elimination Constraint Satisfaction Logics STRIPS Belief Nets Vars + Constraints Decision Nets Var. Elimination Static Sequential Representation Reasoning Technique SLS Markov Chains

Simple Decision Net Early in the morning. Shall I take my umbrella today? (I’ll have to go for a long walk at noon) Relevant Random Variables? CPSC 422, Lecture 2Slide 5

Polices for Umbrella Problem A policy specifies what an agent should do under each circumstance (for each decision, consider the parents of the decision node) In the Umbrella case: D1D1 pD 1 How many policies? One possible Policy CPSC 422, Lecture 2Slide 6

CPSC 422, Lecture 2Slide 7 Value of Information Early in the morning. I listen to the weather forecast, shall I take my umbrella today? (I’ll have to go for a long walk at noon) What would help the agent make a better Umbrella decision?

CPSC 422, Lecture 2Slide 8 Value of Information The value of information of a random variable X for decision D is: the utility of the network with an arc from X to D minus the utility of the network without the arc. Intuitively: The value of information is always It is positive only if the agent changes

CPSC 422, Lecture 2Slide 9 Value of Information (cont.) The value of information provides a bound on how much you should be prepared to pay for a sensor. How much is a perfect weather forecast worth? Original maximum expected utility: Maximum expected utility when we know Weather: Better forecast is worth at most: 77 91

CPSC 422, Lecture 2Slide 10 Value of Information The value of information provides a bound on how much you should be prepared to pay for a sensor. How much is a perfect fire sensor worth? Original maximum expected utility: Maximum expected utility when we know Fire: Perfect fire sensor is worth:

CPSC 422, Lecture 2Slide 11 Value of Control The value of control of a variable X is : the utility of the network when you make X a decision variable minus the utility of the network when X is a random variable. What would help the agent to make an even better Umbrella decision? To maximize its utility. Weather Umbrella Value Rain true Rain false noRain true noRain false

CPSC 422, Lecture 2Slide 12 Value of Control What if we could control the weather? Original maximum expected utility: Maximum expected utility when we control the weather: Value of control of the weather:

CPSC 422, Lecture 2Slide 13 Value of Control What if we control Tampering? Original maximum expected utility: Maximum expected utility when we control the Tampering: Value of control of Tampering: Let’s take a look at the policy Conclusion: do not tamper with fire alarms!

CPSC 422, Lecture 214 Lecture Overview Value of Information and Value of Control Markov Decision Processes (MDPs) Formal Specification and example Define a policy for an MDP

CPSC 422, Lecture 2Slide 15 Combining ideas for Stochastic planning What is a key limitation of decision networks? What is an advantage of Markov models? Represent (and optimize) only a fixed number of decisions The network can extend indefinitely Goal: represent (and optimize) an indefinite sequence of decisions

Slide 16 Infinite horizon problems: process does not stop Indefinite horizon problem: the agent does not know when the process may stop Finite horizon: the process must end at a give time N CPSC 422, Lecture 2 Decision Processes Often an agent needs to go beyond a fixed set of decisions – Examples? Would like to have an ongoing decision process

CPSC 422, Lecture 2Slide 17 Planning in Stochastic Environments Environment Problem Query Planning Deterministic Stochastic Search Arc Consistency Search Value Iteration Var. Elimination Constraint Satisfaction Logics STRIPS Belief Nets Vars + Constraints Decision Nets Markov Decision Processes Var. Elimination Static Sequential Representation Reasoning Technique SLS Markov Chains and HMMs

Slide 18 Recap: Markov Models CPSC 422, Lecture 2

Slide 19 Markov Models Markov Chains Hidden Markov Model Markov Decision Processes (MDPs) CPSC 422, Lecture 2 Partially Observable Markov Decision Processes (POMDPs)

Slide 20 How can we deal with indefinite/infinite Decision processes? We make the same two assumptions we made for…. The action outcome depends only on the current state Let S t be the state at time t … The process is stationary… We also need a more flexible specification for the utility. How? Defined based on a reward/punishment R(s) that the agent receives in each state s CPSC 422, Lecture 2

Slide 21 MDP: formal specification For an MDP you specify: set S of states and set A of actions the process’ dynamics (or transition model) P(S t+1 |S t, A t ) The reward function R(s, a,, s’) describing the reward that the agent receives when it performs action a in state s and ends up in state s’ R(s) is used when the reward depends only on the state s and not on how the agent got there Absorbing/stopping/terminal state CPSC 422, Lecture 2

Slide 22 MDP graphical specification Basically a MDP is a Markov Chain augmented with actions and rewards/values CPSC 422, Lecture 2

Slide 23 When Rewards only depend on the state CPSC 422, Lecture 2

Slide 24 Learning Goals for today’s class You can: Define and compute Value of Information and Value of Control in a decision network Effectively represent indefinite/infinite decision processes with a Markov Decision Process (MDP)

CPSC 422, Lecture 2Slide 25 For Fro: assignment0 – Google Form Read textbook Value of a Policy Value of an Optimal Policy Value Iteration TODO for this Fri