Simulation Selection Problems: Overview of an Economic Analysis Based On Paper By: Stephen E. Chick Noah Gans Presented By: Michael C. Jones MSIM 852.

Slides:



Advertisements
Similar presentations
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Advertisements

Probability Three basic types of probability: Probability as counting
Chapter 6 Sampling and Sampling Distributions
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Why present value leads to better investment decisions than other criteria Yes, there are two paths you can go by, but in the long run there’s still time.
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
Chapter 18 Statistical Decision Theory Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th.
Chapter 21 Statistical Decision Theory
Questions?. Setting a reward function, with and without subgoals Difference between agent and environment AI for games, Roomba Markov Property – Broken.
Chapter 10: Estimating with Confidence
Yazd University, Electrical and Computer Engineering Department Course Title: Machine Learning By: Mohammad Ali Zare Chahooki Bayesian Decision Theory.
1 Chapter 12 Value of Information. 2 Chapter 12, Value of information Learning Objectives: Probability and Perfect Information The Expected Value of Information.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Review: Net Present Value Presentation by: Heather Collins & Michael Maur.
AI – CS364 Uncertainty Management Introduction to Uncertainty Management 21 st September 2006 Dr Bogdan L. Vrusias
Planning under Uncertainty
Statistical Learning: Bayesian and ML COMP155 Sections May 2, 2007.
Jointly Optimal Transmission and Probing Strategies for Multichannel Systems Saswati Sarkar University of Pennsylvania Joint work with Sudipto Guha (Upenn)
1 An Asymptotically Optimal Algorithm for the Max k-Armed Bandit Problem Matthew Streeter & Stephen Smith Carnegie Mellon University NESCAI, April
Network Goodness and its Relation to Probability PDP Class Winter, 2010 January 13, 2010.
Consider the following one-year investment. An amount X is invested in a fund at the beginning of the year. In six months, the fund is worth X/2 at which.
Multi-armed Bandit Problems with Dependent Arms
Optimal rotation age (ORA) Dynamic optimization problem Long discussed in economic literature Shorter rotation – benefit arrives earlier – earlier replanting.
Inductive Reasoning Bayes Rule. Urn problem (1) A B A die throw determines from which urn to select balls. For outcomes 1,2, and 3, balls are picked from.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
1 Monte-Carlo Planning: Policy Improvement Alan Fern.
QA in Finance/ Ch 3 Probability in Finance Probability.
Computational Stochastic Optimization: Bridging communities October 25, 2012 Warren Powell CASTLE Laboratory Princeton University
Search and Planning for Inference and Learning in Computer Vision
Brian Duddy.  Two players, X and Y, are playing a card game- goal is to find optimal strategy for X  X has red ace (A), black ace (A), and red two (2)
Reinforcement Learning Evaluative Feedback and Bandit Problems Subramanian Ramamoorthy School of Informatics 20 January 2012.
* Research is an investigation * It requires the use of the scientific process in order to discover ‘truth’ * Research requires study, observation, comparison.
Chap 20-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 20 Sampling: Additional Topics in Sampling Statistics for Business.
Upper Confidence Trees for Game AI Chahine Koleejan.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
4.05 Implement Accounting Procedures To Track Money Flow And To Determine Financial Status. Describe the nature of income statements.
Probabilistic Results for Mixed Criticality Real-Time Scheduling Bader N. Alahmad Sathish Gopalakrishnan.
Decision Making Under Uncertainty and Risk 1 By Isuru Manawadu B.Sc in Accounting Sp. (USJP), ACA, AFM
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
INFORMS Annual Meeting San Diego 1 HIERARCHICAL KNOWLEDGE GRADIENT FOR SEQUENTIAL SAMPLING Martijn Mes Department of Operational Methods for.
Investment in Human Capital Model-Part I Topic 3 Part III.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Probability 2.0. Independent Events Events can be "Independent", meaning each event is not affected by any other events. Example: Tossing a coin. Each.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Uncertainty Management in Rule-based Expert Systems
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
Making sense of randomness
Once again about the science-policy interface. Open risk management: overview QRAQRA.
The Essence of PDP: Local Processing, Global Outcomes PDP Class January 16, 2013.
Hidden Markov Model Multiarm Bandits: A Methodology for Beam Scheduling in Multitarget Tracking Presented by Shihao Ji Duke University Machine Learning.
Copyright © 2009 Cengage Learning 22.1 Chapter 22 Decision Analysis.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 23 Decision Analysis.
L56 – Discrete Random Variables, Distributions & Expected Values
1 Monte-Carlo Planning: Policy Improvement Alan Fern.
1 Monte-Carlo Planning: Policy Improvement Alan Fern.
Fundamentals of Decision Theory Chapter 16 Mausam (Based on slides of someone from NPS, Maria Fasli)
2.2.1 U NDERSTANDINGMANAGEMENT DECISION MAKING AQA Business 2 M ANAGERS, LEADERSHIP AND DECISION MAKING What decisions do managers make: On a day to day.
Auctions serve the dual purpose of eliciting preferences and allocating resources between competing uses. A less fundamental but more practical reason.
Abdul-Rahman Elshafei – ID  Introduction  SLAT & iSTAT  Multiplet Scoring  Matching Passing Tests  Matching Complex Failures  Multiplet.
+ Chapter 5 Overview 5.1 Introducing Probability 5.2 Combining Events 5.3 Conditional Probability 5.4 Counting Methods 1.
Lesson 4-6 Probability of Compound Events Objectives: To find the probability of independent and dependent events.
Possible actions: up, down, right, left Rewards: – 0.04 if non-terminal state Environment is observable (i.e., agent knows where it is) MDP = “Markov Decision.
PROBABILITY UNLIKELY – LESS ½ LIKELY – MORE ½ CERTAIN - ALL IMPOSSIBLE- CAN’T HAPPEN, 0 EQUAL CHANCE- SAME NUMBER.
Reinforcement Learning Guest Lecturer: Chengxiang Zhai Machine Learning December 6, 2001.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Bayesian Optimization. Problem Formulation Goal  Discover the X that maximizes Y  Global optimization Active experimentation  We can choose which values.
Figure 5: Change in Blackjack Posterior Distributions over Time.
Decisions Under Risk and Uncertainty
Bayesian Statistics and Belief Networks
Presentation transcript:

Simulation Selection Problems: Overview of an Economic Analysis Based On Paper By: Stephen E. Chick Noah Gans Presented By: Michael C. Jones MSIM 852 Fall 2007

Problem Statement: A manager must select a project to implement from a group of potential projects. The manager has a model of each project, and some idea of the statistics of the projects, but does not know the expected value of any project… The manager must select a project, or chose to simulate a project, or chose to abandon all of them and do nothing…

Problem Statement: Running a simulation costs money, and takes time. Recognize that taking time reduces the Net Present Value (NPV) of any project ultimately selected.

Net Present Value: How much is information worth?

Bayes’ Rule "The essence of the Bayesian approach is to provide a mathematical rule explaining how you should change your existing beliefs in the light of new evidence. In other words, it allows scientists to combine new data with their existing knowledge or expertise. The canonical example is to imagine that a precocious newborn observes his first sunset, and wonders whether the sun will rise again or not. He assigns equal prior probabilities to both possible outcomes, and represents this by placing one white and one black marble into a bag. The following day, when the sun rises, the child places another white marble in the bag. The probability that a marble plucked randomly from the bag will be white (ie, the child's degree of belief in future sunrises) has thus gone from a half to two-thirds. After sunrise the next day, the child adds another white marble, and the probability (and thus the degree of belief) goes from two-thirds to three-quarters. And so on. Gradually, the initial belief that the sun is just as likely as not to rise each morning is modified to become a near-certainty that the sun will always rise." * *”In Praise of Bayes,” The Economist, 30 Sept 2000

Bayes’ Rule An example will help…

Bayes’ Rule Suppose you have two bags of marbles to select from. One bag (a 1 ) has 75% white marbles and 25% black marbles, the other bag (a 2 ) has 25% white marbles and 75% black marbles. You pick a bag, but do not know which one. From the bag, you select 5 marbles (with replacement). They are 4 white and 1 black marbles. Call the experiment B. What bag do you have? What is P(a 1 |B)? What is P(a 2 |B)?

Bayes’ Rule but: and: so:

Bayes’ Rule but: and: so:

Multi-armed Bandit Problem A gambler has a slot machine with several arms (or several slot machines to select from) but does not know the expected payout. How can he maximize his probability of playing the machine with the highest payout?

Multi-armed Bandit Problem Classic Approach –Epsilon First –Epsilon Greedy –Epsilon Decreasing Indifference Zones… Gittins Index…

Multi-armed Bandit Problem Gittins Index… –Gittins found a dynamic allocation index based on Baye’s rule. –Does not follow Bayes directly. (Consider what could happen after the first pull if he followed Bayes exactly…) –Weights the results by the “regret” cost of incorrect selection, discounted to the horizon.

Gittins’ Index: Gittin’s Index is the value of the game: –For several machines, n= 1, 2, …N –machine n is in state i. If machine n is selected, a reward,, is earned. Machine n then transitions to state j, with some known probability.

Returning to the Manager: Gittin’s Index is difficult (and computationally expensive) to calculate. The Index may be approximated by the Optimal Expected Discounted Reward:

Returning to the Manager: The Index may be approximated by the Optimal Expected Discounted Reward. Select the most profitable of: –Do nothing. E[] = 0 –Simulate a project. E[] = –Implement a project now. E[] =

Procedure

Summary Strengths: –Provides a tools for accounting for cost (both opportunity and direct) of simulation. –May be automated so the business manager can apply the tool. Weaknesses: –Limiting assumptions (Known variance but unknown mean?) –Mathematically abstract.

Future Not applicable directly The authors state that the paper presents more questions than it answers. May provide stimulation for future research.

Future Not applicable directly The authors state that the paper presents more questions than it answers. May provide stimulation for future research. “If I have seen a little further it is by standing on the shoulders of Giants." Newton, 1676