The AutoSimOA Project Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC 07 A 3 year, EPSRC funded project in collaboration with SIMUL8.

Slides:



Advertisements
Similar presentations
Introduction to parameter optimization
Advertisements

Automating the Analysis of Simulation Output Data Katy Hoad Stewart Robinson, Ruth Davies, Mark Elder Funded by EPSRC and SIMUL8.
Chap 8-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 8 Estimation: Single Population Statistics for Business and Economics.
Chapter 11- Confidence Intervals for Univariate Data Math 22 Introductory Statistics.
Output analyses for single system
The Comparison of the Software Cost Estimating Methods
1 Statistical Inference H Plan: –Discuss statistical methods in simulations –Define concepts and terminology –Traditional approaches: u Hypothesis testing.
Output Data Analysis. How to analyze simulation data? simulation –computer based statistical sampling experiment –estimates are just particular realizations.
Automating estimation of warm-up length Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC08 The AutoSimOA Project A 3 year, EPSRC funded.
Chapter 12 Simple Regression
Automating estimation of warm-up length Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School Simulation Workshop - April 2008 The AutoSimOA.
Simulation Modeling and Analysis Session 12 Comparing Alternative System Designs.
Automated Analysis of Simulation Output Data and the AutoSimOA Project
Automating the Analysis of Simulation Output Data Katy Hoad, Stewart Robinson, Ruth Davies SSIG Meeting, 24th October 2007
Simulation.
Evaluating Hypotheses
Classification of Discrete Event Simulation Models and Output Data: Creating a Sufficient Model Set. Katy Hoad Stewart Robinson,
The AutoSimOA Project Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School OR49 Sept 07 A 3 year, EPSRC funded project in collaboration with.
Automating the Analysis of Simulation Output Data Stewart Robinson, Katy Hoad, Ruth Davies OR48, September 2006.
Lecture 9: One Way ANOVA Between Subjects
Automating the Analysis of Simulation Output Data Katy Hoad Stewart Robinson, Ruth Davies, Mark Elder Funded by EPSRC and SIMUL8.
EEM332 Design of Experiments En. Mohd Nazri Mahmud
Automating the Analysis of Simulation Output Data Stewart Robinson, Katy Hoad, Ruth Davies ORGS Meeting, 4th October 2007
SIMULATION. Simulation Definition of Simulation Simulation Methodology Proposing a New Experiment Considerations When Using Computer Models Types of Simulations.
1 Simulation Modeling and Analysis Output Analysis.
191 Drawing Statistical Inference from Simulation Runs......the "fun" stuff!
PSY 307 – Statistics for the Behavioral Sciences Chapter 7 – Regression.
Automating The Selection of a Simulation Warm-up Period Stewart Robinson, Katy Hoad, Ruth Davies Warwick Business School Cardiff University - October 2008.
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
AutoSimOA : A Framework for Automated Analysis of Simulation Output Stewart Robinson Katy Hoad, Ruth Davies Funded by.
Scot Exec Course Nov/Dec 04 Ambitious title? Confidence intervals, design effects and significance tests for surveys. How to calculate sample numbers when.
INFERENTIAL STATISTICS – Samples are only estimates of the population – Sample statistics will be slightly off from the true values of its population’s.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 12-1 Chapter 12 Simple Linear Regression Statistics for Managers Using.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Introduction to Statistical Inferences
Determining Sample Size
Verification & Validation
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Basic Business Statistics 11 th Edition.
Confidence Interval Estimation
Review of Chapters 1- 5 We review some important themes from the first 5 chapters 1.Introduction Statistics- Set of methods for collecting/analyzing data.
MS 305 Recitation 11 Output Analysis I
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Estimating a Population Mean: σ Known 7-3, pg 355.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Brian Macpherson Ph.D, Professor of Statistics, University of Manitoba Tom Bingham Statistician, The Boeing Company.
Lecture 2 Forestry 3218 Lecture 2 Statistical Methods Avery and Burkhart, Chapter 2 Forest Mensuration II Avery and Burkhart, Chapter 2.
Measures of central tendency are statistics that express the most typical or average scores in a distribution These measures are: The Mode The Median.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Lecture 7 Dustin Lueker. 2  Point Estimate ◦ A single number that is the best guess for the parameter  Sample mean is usually at good guess for the.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
Chapter 8: Confidence Intervals based on a Single Sample
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
1 Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Example: In a recent poll, 70% of 1501 randomly selected adults said they believed.
Quality Control: Analysis Of Data Pawan Angra MS Division of Laboratory Systems Public Health Practice Program Office Centers for Disease Control and.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Business Statistics: A First Course 5 th Edition.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 15: Sample size and Power Marshall University Genomics.
Stats Methods at IC Lecture 3: Regression.
Slides by JOHN LOUCKS St. Edward’s University.
ESTIMATION.
Confidence Interval Estimation
Sampling Distributions and Estimation
Al-Imam Mohammad Ibn Saud University Large-Sample Estimation Theory
Chapter 8: Inference for Proportions
CPSC 531: System Modeling and Simulation
Chapter 10 Verification and Validation of Simulation Models
Statistical Methods Carey Williamson Department of Computer Science
Confidence Interval Estimation
Carey Williamson Department of Computer Science University of Calgary
Objectives 6.1 Estimating with confidence Statistical confidence
Presentation transcript:

The AutoSimOA Project Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC 07 A 3 year, EPSRC funded project in collaboration with SIMUL8 Corporation.

Objective To provide an easy to use method, that can be incorporated into existing simulation software, that enables practitioners to obtain results of a specified accuracy from their discrete event simulation model. (Only looking at analysis of a single scenario)

OUTLINE Introduction Methods in literature Our Algorithm Test Methodology & Results Discussion & Summary

Underlying Assumptions  Any warm-up problems already dealt with.  Run length (m) decided upon.  Modeller decided to use multiple replications to obtain better estimate of mean performance. Response measure of interest summary statistic from each replication Perform N replications

QUESTION IS… How many replications are needed? Limiting factors: computing time and expense. 4 main methods found in the literature for choosing the number of replications N to perform.

1. Rule of Thumb (Law & McComas 1990) Run at least 3 to 5 replications. Advantage: Very simple. Disadvantage: Does not use characteristics of model output. No measured precision level.

2. Simple Graphical Method (Robinson 2004) Advantages: Simple Uses output of interest in decision. Disadvantages: Subjective No measured precision level.

3. Confidence Interval Method (Robinson 2004, Law 2007, Banks et al. 2005). Advantages: Uses statistical inference to determine N. Uses output of interest in decision. Provides specified precision. Disadvantage: Many simulation users do not have the skills to apply approach.

4. Prediction Formula (Banks et al. 2005) Decide size of error ε that can be can tolerated. Run ≥ 2 replications - estimate variance s 2. Solve to predict N. Check desired precision achieved – if not recalculate N with new estimate of variance. Advantages: Uses statistical inference to determine N. Uses output of interest in decision. Provides specified precision. Disadvantage: Can be very inaccurate especially for small number of replications.

AUTOMATE Confidence Interval Method: Algorithm interacts with simulation model sequentially.

is the student t value for n-1 df and a significance of 1-α, s n is the estimate of the standard deviation, calculated using results X i (i = 1 to n) of the n current replications. Where n is the current number of replications carried out, We define the precision, d n, as the ½ width of the Confidence Interval expressed as a percentage of the cumulative mean: is the cumulative mean, ALGORITHM DEFINITIONS

Stopping Criteria Simplest method: Stop when d n 1st found to be ≤ desired precision, d required. Recommend that number of replications, Nsol, to user. Problem: Data series could prematurely converge, by chance, to incorrect estimate of the mean, with precision d required, then diverge again. ‘Look-ahead’ procedure: When d n 1st found to be ≤ d required, algorithm performs set number of extra replications, to check that precision remains ≤ d required.

‘Look-ahead’ procedure kLimit = ‘look ahead’ value. Actual number of replications checked ahead is Relates ‘look ahead’ period length with current value of n.

Nsol Nsol + f(kLimit) f(kLimit) Precision ≤ 5% 95% confidence limits Cumulative mean, Replication Algorithm

Precision≤ 5% Precision> 5% Precision ≤ 5% f(kLimit) Nsol 2 Nsol 2 + f(kLimit) Nsol 1

24 artificial data sets: Left skewed, symmetric, right skewed; Varying values of relative st.dev (st.dev/mean). 100 sequences of 2000 data values. 8 real models selected. Different lengths of ‘look ahead’ period tested: kLimit values = 0 (i.e. no ‘look ahead’), 5, 10, 25. d required value kept constant at 5%. TESTING METHODOLOGY

5 performance measures 1.Coverage of the true mean 2.Bias 3.Absolute Bias 4.Average Nsol value 5.Comparison of 4. with Theoretical Nsol value For real models: ‘true’ mean & variance values - estimated from whole sets of output data (3000 to data points).

Results Nsol values for individual algorithm runs are very variable. Average Nsol values for 100 runs per model close to the theoretical values of Nsol. Normality assumption appears robust. Using a ‘look ahead’ period improves performance of the algorithm.

Mean bias significantly different to zero Failed in coverage of true mean Mean est. Nsol significantly different to theoretical Nsol (>3) No ‘look- ahead’ period Proportion of Artificial models 4/242/249/18 Proportion of Real models 1/8 3/5 kLimit = 5 Proportion of Artificial models 1/2401/18 Proportion of Real models 000

% decrease in absolute mean bias kLimit = 0 to kLimit = 5 kLimit = 5 to kLimit = 10 kLimit = 10 to kLimit = 25 Artificial Models 8.76%0.07%0.26% Real Models 10.45%0.14%0.33% Impact of different look ahead periods on performance of algorithm

Number of times the Nsol value changes (out of 100 runs of the algorithm per model) because of the lengthening of the ‘look ahead’ period. Model ID kLimit = 0 to kLimit = 5 kLimit = 5 to kLimit = 10 kLimit = 10 to kLimit = 25 R1000 R3200 R52401 R82441 A53013 A62663 A15100 A A A243700

Model ID kLimitNsolTheoretical Nsol (approx) Mean estimate significantly different to the true mean? A904112Yes 5120No A Yes 5718No R70310Yes 58No R4036Yes 57No R80345Yes 546No Examples of changes in Nsol & improvement in estimate of true mean

DISCUSSION kLimit default value set to 5. Initial number of replications set to 3. Multiple response variables - Algorithm run with each response - use maximum estimated value for Nsol. Different scenarios - advisable to repeat algorithm every few scenarios to check that precision has not degraded significantly. Implementation into Simul8 simulation package.

SUMMARY Selection and automation of Confidence Interval Method for estimating the number of replications to be run in a simulation. Algorithm created with ‘look ahead’ period - efficient and performs well on wide selection of artificial and real model output. ‘Black box’ - fully automated and does not require user intervention.

ACKNOWLEDGMENTS This work is part of the Automating Simulation Output Analysis (AutoSimOA) project ( that is funded by the UK Engineering and Physical Sciences Research Council (EP/D033640/1). The work is being carried out in collaboration with SIMUL8 Corporation, who are also providing sponsorship for the project. Katy Hoad, Stewart Robinson, Ruth Davies Warwick Business School WSC 07