Training Course 2009 – NWP-PR: Ensemble Verification II 1/33 Ensemble Verification II Renate Hagedorn European Centre for Medium-Range Weather Forecasts.

Slides:



Advertisements
Similar presentations
How Well Forecast Were the 2004 and 2005 Atlantic and U. S
Advertisements

Feichter_DPG-SYKL03_Bild-01. Feichter_DPG-SYKL03_Bild-02.
Climate Modeling LaboratoryMEASNC State University An Extended Procedure for Implementing the Relative Operating Characteristic Graphical Method Robert.
Copyright © 2008 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics, 9e Managerial Economics Thomas Maurice.
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Chapter 1 The Study of Body Function Image PowerPoint
1 Copyright © 2013 Elsevier Inc. All rights reserved. Appendix 01.
STATISTICS HYPOTHESES TEST (III) Nonparametric Goodness-of-fit (GOF) tests Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering.
UNITED NATIONS Shipment Details Report – January 2006.
GEO Energy Management Meeting WMO, August 2006 Earth Observations and Energy Management Expert Meeting Renate Hagedorn European Centre for Medium-Range.
Aims and Requirements for Ensemble Forecasting By T.N.Palmer ECMWF.
THOR Annual Meeting - Bergen 9-11 November /25 On the impact of initial conditions relative to external forcing on the skill of decadal predictions:
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Extended range forecasts at MeteoSwiss: User experience.
ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF.
Slide 1ECMWF forecast User Meeting -- Reading, June 2006 Verification of weather parameters Anna Ghelli, ECMWF.
User Meeting 15 June 2005 Monthly Forecasting Frederic Vitart ECMWF, Reading, UK.
Severe Weather Forecasts
Sub-seasonal to seasonal prediction David Anderson.
Exit a Customer Chapter 8. Exit a Customer 8-2 Objectives Perform exit summary process consisting of the following steps: Review service records Close.
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
Year 6 mental test 5 second questions
CS1512 Foundations of Computing Science 2 Week 3 (CSD week 32) Probability © J R W Hunter, 2006, K van Deemter 2007.
1 Correlation and Simple Regression. 2 Introduction Interested in the relationships between variables. What will happen to one variable if another is.
Evaluating Provider Reliability in Risk-aware Grid Brokering Iain Gourlay.
© 2002 Prentice-Hall, Inc.Chap 17-1 Basic Business Statistics (8 th Edition) Chapter 17 Decision Making.
Chapter 7 Sampling and Sampling Distributions
Break Time Remaining 10:00.
(This presentation may be used for instructional purposes)
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
EU Market Situation for Eggs and Poultry Management Committee 21 June 2012.
5-1 Chapter 5 Theory & Problems of Probability & Statistics Murray R. Spiegel Sampling Theory.
1 Slides revised The overwhelming majority of samples of n from a population of N can stand-in for the population.
Discrete Event (time) Simulation Kenneth.
Oil & Gas Final Sample Analysis April 27, Background Information TXU ED provided a list of ESI IDs with SIC codes indicating Oil & Gas (8,583)
15. Oktober Oktober Oktober 2012.
Meteorological Training Course, 20 March /25 Using Combined Prediction Systems (CPS) for wind energy applications European Centre for Medium-Range.
Factor P 16 8(8-5ab) 4(d² + 4) 3rs(2r – s) 15cd(1 + 2cd) 8(4a² + 3b²)
1..
What is a good ensemble forecast? Chris Ferro University of Exeter, UK With thanks to Tom Fricker, Keith Mitchell, Stefan Siegert, David Stephenson, Robin.
Lecture 3 Validity of screening and diagnostic tests
© 2012 National Heart Foundation of Australia. Slide 2.
What is a good ensemble forecast? Chris Ferro University of Exeter, UK With thanks to Tom Fricker, Keith Mitchell, Stefan Siegert, David Stephenson, Robin.
Statistical Analysis SC504/HS927 Spring Term 2008
1 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt 10 pt 15 pt 20 pt 25 pt 5 pt Synthetic.
Chapter 8: Introduction to Hypothesis Testing. 2 Hypothesis Testing An inferential procedure that uses sample data to evaluate the credibility of a hypothesis.
Model and Relationships 6 M 1 M M M M M M M M M M M M M M M M
Evaluation of precision and accuracy of a measurement
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
Chapter Thirteen The One-Way Analysis of Variance.
Clock will move after 1 minute
Intracellular Compartments and Transport
PSSA Preparation.
Essential Cell Biology
Multiple Regression and Model Building
Murach’s OS/390 and z/OS JCLChapter 16, Slide 1 © 2002, Mike Murach & Associates, Inc.
4/4/2015Slide 1 SOLVING THE PROBLEM A one-sample t-test of a population mean requires that the variable be quantitative. A one-sample test of a population.
Key Concepts and Skills
Guidance of the WMO Commission for CIimatology on verification of operational seasonal forecasts Ernesto Rodríguez Camino AEMET (Thanks to S. Mason, C.
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Creating probability forecasts of binary events from ensemble predictions and prior information - A comparison of methods Cristina Primo Institute Pierre.
© Crown copyright Met Office Operational OpenRoad verification Presented by Robert Coulson.
Verification of ensembles Courtesy of Barbara Brown Acknowledgments: Tom Hamill, Laurence Wilson, Tressa Fowler Copyright UCAR 2012, all rights reserved.
ECMWF Training Course Reading, 25 April 2006 EPS Diagnostic Tools Renate Hagedorn European Centre for Medium-Range Weather Forecasts.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
EUMETCAL NWP-course 2007: The Concept of Ensemble Forecasting Renate Hagedorn European Centre for Medium-Range Weather Forecasts The General Concept of.
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Common verification methods for ensemble forecasts
Verifying and interpreting ensemble products
the performance of weather forecasts
Presentation transcript:

Training Course 2009 – NWP-PR: Ensemble Verification II 1/33 Ensemble Verification II Renate Hagedorn European Centre for Medium-Range Weather Forecasts

Training Course 2009 – NWP-PR: Ensemble Verification II 2/33 Assessing the quality of a forecast system Characteristics of a forecast system: Consistency: Do the observations statistically belong to the distributions of the forecast ensembles? (consistent degree of ensemble dispersion) Reliability: Can I trust the probabilities to mean what they say? Sharpness: How much do the forecasts differ from the climatological mean probabilities of the event? Resolution: How much do the forecasts differ from the climatological mean probabilities of the even, and the systems gets it right? Skill: Are the forecasts better than my reference system (chance, climatology, persistence,…)? Reliability Diagram Rank Histogram Brier Skill Score

Training Course 2009 – NWP-PR: Ensemble Verification II 3/33 Brier Score -> Ranked Probability Score f(y) Brier Score used for two category (yes/no) situations (e.g. T > 15 o C) RPS takes into account ordered nature of variable (extreme errors) F(y) 1

Training Course 2009 – NWP-PR: Ensemble Verification II 4/33 Ranked Probability Score category f(y) category F(y) 1 PDF CDF

Training Course 2009 – NWP-PR: Ensemble Verification II 5/33 Ranked Probability Score category f(y) category F(y) 1 PDF CDF

Training Course 2009 – NWP-PR: Ensemble Verification II 6/33 Ranked Probability Score Measures the quadratic distance between forecast and verification probabilities for several probability categories k It is the average Brier score across the range of the variable Ranked Probability Skill Score (RPSS) is a measure for skill relative to a reference forecast Emphasizes accuracy by penalizing large errors more than near misses Rewards sharp forecast if it is accurate

Training Course 2009 – NWP-PR: Ensemble Verification II 7/33 Ranked Probability Score category f(y) PDF RPS=0.01 sharp & accurate category f(y) PDF RPS=0.15 sharp, but biased category f(y) PDF RPS=0.05 not very sharp, slightly biased category f(y) PDF RPS=0.08 accurate, but not sharp climatology

Training Course 2009 – NWP-PR: Ensemble Verification II 8/33 Definition of a proper score Consistency with your true belief is one of the characteristics of a good forecast Some scoring rules encourage forecasters to be inconsistent, e.g. some scores give better results when a forecast closer to climatology is issued rather than the actual forecast (e.g. reliability) Scoring rule is strictly proper when the best scores are obtained if and only if the forecasts correspond with the forecasters judgement (true belief) Examples of proper scores are the Brier Score or RPS or Ignorance Score

Training Course 2009 – NWP-PR: Ensemble Verification II 9/33 Ignorance Score N: number of observation-forecast pairs I: Quantiles o n,i : observation probability (0 or 1) p n,i : forecast probability Minimum only when p n,i = o n,i = proper score The lower/higher the IGN the better/worse the forecast system See Roulston & Smith, 2002

Training Course 2009 – NWP-PR: Ensemble Verification II 10/33 Brier Score vs. Ignorance Score - log(p) (p – o) 2

Training Course 2009 – NWP-PR: Ensemble Verification II 11/33 Why Probabilities? Open air restaurant scenario: open additional tables: £20 extra cost, £100 extra income (if T>24ºC) weather forecast: 30% probability for T>24ºC what would you do? Test the system for 100 days: 30 x T>24ºC -> 30 x (100 – 20) = x T 70 x ( 0 – 20) = Employing extra waiter (spending £20) is beneficial when probability for T>24 ºC is greater 20% The higher/lower the cost loss ratio, the higher/lower probabilities are needed in order to benefit from action on forecast

Training Course 2009 – NWP-PR: Ensemble Verification II 12/33 Benefits for different users - decision making A user (or decision maker) is sensitive to a specific weather event The user has a choice of two actions: do nothing and risk a potential loss L if weather event occurs take preventative action at a cost C to protect against loss L Decision-making depends on available information: no FC information: either always take action or never take action deterministic FC: act when adverse weather predicted probability FC: act when probability of specific event exceeds a certain threshold (this threshold depends on the user) Value V of a forecast: savings made by using the forecast, normalized so that V = 1 for perfect forecast V = 0 for forecast not better than climatology Ref: D. Richardson, 2000, QJRMS

Training Course 2009 – NWP-PR: Ensemble Verification II 13/33 Decision making: the cost-loss model Climate information – expense: Always use forecast – expense: Perfect forecast – expense: Value: Event occurs YesNo Action taken YesCC NoL0 Event occurs YesNo Event forecast Yesab Nocd o1-o Potential costs Fraction of occurences

Training Course 2009 – NWP-PR: Ensemble Verification II 14/33 with: α = C/L H = a/(a+c) F = b/(b+d) ō = a+c Decision making: the cost-loss model For given weather event and FC system: ō, H and F are fixed value depends on C/L max if: C/L = ō V max = H-F Northern Extra-Tropics (winter 01/02) D+5 deterministic FC > 1mm precip

Training Course 2009 – NWP-PR: Ensemble Verification II 15/33 Potential economic value Northern Extra-Tropics (winter 01/02) D+5 FC > 1mm precipitation deterministicEPS p = 0.2 p = 0.5 p = 0.8

Training Course 2009 – NWP-PR: Ensemble Verification II 16/33 Potential economic value EPS: each user chooses the most appropriate probability threshold Control Results based on simple cost/loss models have indicated that EPS probabilistic forecasts have a higher value than single deterministic forecasts EPS Northern Extra-Tropics (winter 01/02) D+5 FC > 1mm precipitation

Training Course 2009 – NWP-PR: Ensemble Verification II 17/33 Potential economic value Northern Extra-Tropics (winter 01/02) D+5 FC > 20mm precipitation BSS = 0.06 (measure of overall value for all possible users) ROCSS = 0.65 (closely linked to V max )

Training Course 2009 – NWP-PR: Ensemble Verification II 18/33 Variation of value with higher resolution Relative improvement of higher resolution for different C/L ratios (57 winter cases, 850hPa temperature, positive anomalies) Forecast day

Training Course 2009 – NWP-PR: Ensemble Verification II 19/33 Variation of value with ensemble size 10 ensemble members 50 ensemble members Underlying distribution (large ensemble limit) Ref: D. Richardson, 2006, in Palmer & Hagedorn

Training Course 2009 – NWP-PR: Ensemble Verification II 20/33 Weather Roulette The funding agency of a weather forecast centre believes that the forecasts are useless and not better than climatology! The Director of the weather centre believes that their forecasts are more worth than a climatological forecast! She challenges the funding agency in saying: I bet, I can make more money with our forecasts than you can make with a climatological forecast!

Training Course 2009 – NWP-PR: Ensemble Verification II 21/33 Weather Roulette Both parties, the funding agency (A) and the Director (D), agree that both of them open a weather roulette casino, and that both of them spend each day 1 k of their own budget in the casino of the other party A & D use their favourite forecast to (i) set the odds of their own casino and (ii) distribute their money in the other casino A sets the odds of its casino and distributes the stake according to climatology D sets the odds of her casino and distributes her stake according to her forecast They agree to bet on the 2m-temperature at London-Heathrow being well-below, below, normal, above, or well-above the long-term climatological average (5 possible categories)

Training Course 2009 – NWP-PR: Ensemble Verification II 22/33 Weather Roulette Verification bin o OBS EPS DET

Training Course 2009 – NWP-PR: Ensemble Verification II 23/33 Odds in casino A: casino D: with: i=1,…,N: possible outcomes p A (i): As probability of the i th outcome p D (i): Ds probability of the i th outcome Weather Roulette Stakes of A: of D: with: c = available capital to be distributed every day Return for A: Return for D: with: v = verifying outcome

Training Course 2009 – NWP-PR: Ensemble Verification II 24/33 Weather Roulette D gets her return r D from A, but has to payout r A to A D can increase the weather centres budget if: = 2.5 – 0.4 = 2.1

Training Course 2009 – NWP-PR: Ensemble Verification II 25/33 Weather Roulette: LHR T2m, D+3 Verification bin EPS Climatology probability of verification bin accumulated winnings for EPS

Training Course 2009 – NWP-PR: Ensemble Verification II 26/33 Weather Roulette: LHR T2m, D+10 Verification bin EPS probability of verification bin accumulated winnings for EPS EPS Climatology

Training Course 2009 – NWP-PR: Ensemble Verification II 27/33 Weather Roulette Adam, working in the Model Division believes that the ECMWF deterministic high- resolution model is the best forecast system in the world! Eve, working in the Probability Forecasting Division believes that the ECMWF EPS is the best forecast system in the world! Eve challenges Adam and says: I bet, I can make more money with my EPS forecasts than you can make with your high- resolution deterministic forecasts!

Training Course 2009 – NWP-PR: Ensemble Verification II 28/33 Dressing The idea: Find an appropriate dressing kernel from past performance (the smaller/greater past error the smaller/greater g_sdev) P exp = [0.0, 1.0, 0.0] [0.25, 0.70, 0.05] /3 p(v q) = Rank(q) – 1/3 (nens+1) + 1/3 p(ver) = = – 1/3 2 – 1/ p(bin) = p sum (bin)/p total

Training Course 2009 – NWP-PR: Ensemble Verification II 29/33 Weather Roulette: LHR T2m, D+3 Verification bin EPS DET probability of verification bin accumulated winnings for EPS EPS (dressed) DET (dressed)

Training Course 2009 – NWP-PR: Ensemble Verification II 30/33 Weather Roulette: LHR T2m, D+10 Verification bin EPS DET probability of verification bin accumulated winnings for EPS EPS (dressed) DET (dressed)

Training Course 2009 – NWP-PR: Ensemble Verification II 31/33 Weather Roulette: 100 stations, MAM 2006 DET vs. CLI EPS vs. CLI EPS vs. DET average daily capital growth

Training Course 2009 – NWP-PR: Ensemble Verification II 32/33 Summary II Different users are sensitive to different weather events They have different cost/loss ratios They have different probability thresholds for their decision- making process Simple cost/loss models indicate that probabilistic forecast systems have a higher potential economic value than deterministic forecasts The relative improvement of increasing model resolution or ensemble size depends on the individual users C/L The weather roulette diagnostic is a useful tool to demonstrate the real benefit of using the EPS

Training Course 2009 – NWP-PR: Ensemble Verification II 33/33 References and further reading Katz, R. W. and A.H. Murphy, 1997: Economic value of weather and climate forecasting. Cambridge University Press, pp Palmer, T. and R. Hagedorn (editors), 2006: Predictability of weather and climate. Cambridge University Press, pp.702 Jolliffe, I.T. and D.B. Stephenson, 2003: Forecast Verification. A Practitioners Guide in Atmospheric Science. Wiley, pp. 240 Wilks, D. S., 2006: Statistical methods in the atmospheric sciences. 2 nd ed. Academic Press, pp.627 Hagedorn, R. and L.A. Smith, 2009: Communicating the value of probabilistic forecasts with weather roulette. Meteorological Applications, in press. Richardson, D. S., Skill and relative economic value of the ECMWF Ensemble Prediction System. Q. J. R. Meteorol. Soc., 126, Hamill, T., 2001: Interpretation of Rank Histograms for Verifying Ensemble Forecasts. Monthly Weather Review, 129, Roulston, M.S. and L.A.Smith, Evaluating probabilistic forecasts using information theory. Monthly Weather Review, 130,