Guidance of the WMO Commission for CIimatology on verification of operational seasonal forecasts Ernesto Rodríguez Camino AEMET (Thanks to S. Mason, C.

Slides:



Advertisements
Similar presentations
ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF.
Advertisements

1 Verification Continued… Holly C. Hartmann Department of Hydrology and Water Resources University of Arizona RFC Verification Workshop,
Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
WA Landman. The RCOF concept  Towards the end of 1997 two important climate-related events occurred The record-breaking El Niño episode The holding of.
14 May 2001QPF Verification Workshop Verification of Probability Forecasts at Points WMO QPF Verification Workshop Prague, Czech Republic May 2001.
Wrap-up. Predicting atmospheric circulation DJF NAO Met Office operational seasonal forecasts with HadGEM3H N216L85O(0.25) with initial conditions from.
1 of Introduction to Forecasts and Verification.
Statistical Issues in Research Planning and Evaluation
Verification of probability and ensemble forecasts
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Bookmaker or Forecaster? By Philip Johnson. Jersey Meteorological Department.
Introduction to Probability and Probabilistic Forecasting L i n k i n g S c i e n c e t o S o c i e t y Simon Mason International Research Institute for.
Gridded OCF Probabilistic Forecasting For Australia For more information please contact © Commonwealth of Australia 2011 Shaun Cooper.
Class 2: Tues., Sept. 14th Correlation (2.2) Introduction to Measurement Theory: –Reliability of measurements and correlation –Example that demonstrates.
Application of seasonal climate forecasts to predict regional scale crop yields in South Africa Trevor Lumsden and Roland Schulze School of Bioresources.
Introduction to Seasonal Climate Prediction Liqiang Sun International Research Institute for Climate and Society (IRI)
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Software Project Management Lecture # 8. Outline Chapter 25 – Risk Management  What is Risk Management  Risk Management Strategies  Software Risks.
OSSE School Improvement Data Workshop Workshop #1 January 30, 2015 Office of the State Superintendent of Education.
Verification of ensembles Courtesy of Barbara Brown Acknowledgments: Tom Hamill, Laurence Wilson, Tressa Fowler Copyright UCAR 2012, all rights reserved.
Classroom Assessments Checklists, Rating Scales, and Rubrics
4IWVM - Tutorial Session - June 2009 Verification of categorical predictands Anna Ghelli ECMWF.
Chapter 1: Research Methods
Verification of the Cooperative Institute for Precipitation Systems‘ Analog Guidance Probabilistic Products Chad M. Gravelle and Dr. Charles E. Graves.
How can LAMEPS * help you to make a better forecast for extreme weather Henrik Feddersen, DMI * LAMEPS =Limited-Area Model Ensemble Prediction.
Verification methods - towards a user oriented verification WG5.
Patterns of Event Causality Suggest More Effective Corrective Actions Abstract: The Occurrence Reporting and Processing System (ORPS) has used a consistent.
Heidke Skill Score (for deterministic categorical forecasts) Heidke score = Example: Suppose for OND 1997, rainfall forecasts are made for 15 stations.
Chapter 6 USING PROBABILITY TO MAKE DECISIONS ABOUT DATA.
Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
1 Psych 5500/6500 The t Test for a Single Group Mean (Part 1): Two-tail Tests & Confidence Intervals Fall, 2008.
Statistics - methodology for collecting, analyzing, interpreting and drawing conclusions from collected data Anastasia Kadina GM presentation 6/15/2015.
Forecasting in CPT Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
61 st IHC, New Orleans, LA Verification of the Monte Carlo Tropical Cyclone Wind Speed Probabilities: A Joint Hurricane Testbed Project Update John A.
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
Can we distinguish wet years from dry years? Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand,
A Superior Alternative to the Modified Heidke Skill Score for Verification of Categorical Versions of CPC Outlooks Bob Livezey Climate Services Division/OCWWS/NWS.
© Crown copyright Met Office Standard Verification System for Long-range Forecasts (SVSLRF) Richard Graham, Met Office Hadley Centre. With acknowledgements.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
1 Probabilistic Forecast Verification Allen Bradley IIHR Hydroscience & Engineering The University of Iowa RFC Verification Workshop 16 August 2007 Salt.
Probabilistic Forecasts of Extreme Precipitation Events for the U.S. Hazards Assessment Kenneth Pelman 32 nd Climate Diagnostics Workshop Tallahassee,
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Common verification methods for ensemble forecasts
Science plan S2S sub-project on verification. Objectives Recommend verification metrics and datasets for assessing forecast quality of S2S forecasts Provide.
Testing. Psychological Tests  Tests abilities, interests, creativity, personality, behavior  Must be standardized, reliable, and valid  Timing, instructions,
Details for Today: DATE:13 th January 2005 BY:Mark Cresswell FOLLOWED BY:Practical Dynamical Forecasting 69EG3137 – Impacts & Models of Climate Change.
Verification methods - towards a user oriented verification The verification group.
Status / Progress Report for GPC Tokyo Takayuki Tokuhiro Climate Prediction Division, JMA
Beginners statistics Assoc Prof Terry Haines. 5 simple steps 1.Understand the type of measurement you are dealing with 2.Understand the type of question.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Extension: How could researchers use a more powerful measure of analysis? Why do you think that researchers do not just rely on descriptive statistics.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 21 More About Tests and Intervals.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Addressing the environmental impact of salt use on roads
Verifying and interpreting ensemble products
Classroom Assessments Checklists, Rating Scales, and Rubrics
Climate Predictability Tool (CPT)
Binary Forecasts and Observations
IPET-OPSLS: Activities and Achievements (after April 2016 Beijing Meeting) Arun Kumar Chair, CBS/CCL IPET-OPSLS Climate Prediction Center, USA
Probabilistic forecasts
Caio Coelho (Joint CBS/CCl IPET-OPSLS Co-chair) CPTEC/INPE, Brazil
COSMO-LEPS Verification
How good (or bad) are seasonal forecasts?
Can we distinguish wet years from dry years?
Forecast system development activities
Seasonal Forecasting Using the Climate Predictability Tool
Verification of SPE Probability Forecasts at SEPC
Seasonal Forecasting Using the Climate Predictability Tool
Short Range Ensemble Prediction System Verification over Greece
Presentation transcript:

Guidance of the WMO Commission for CIimatology on verification of operational seasonal forecasts Ernesto Rodríguez Camino AEMET (Thanks to S. Mason, C. Coelho, C. Santos, E. Sanchez)

Forecasts possess no intrinsic value. They acquire value through their ability to influence the decisions made by users of the forecasts. (Murphy 1993)

Why verify operational seasonal forecasts? Does a new system improve the current one? Is the cost of the forecast justified? Is it a good idea to use (or pay for) the forecast? If so, how can they best used? All operational forecast should be accompanied by readily available information on the quality of forecast (minimum set of diagnostics)

Example: ECMWF Unfortunately, this is not the case for most RCOFs!

Motivation There is quite a lot of experience veryfing probabilistic outputs of seasonal models. Complement the Standardized Verification System for LRFs (SVSLRF) for GPC products.

Example verification seasonal forecasts from GCMs: RPSS

Motivation There is quite a lot of experience veryfing probabilistic outputs of seasonal models. Complement the Standardized Verification System for LRFs (SVSLRF) for GPC products. So far most RCOFs are limited their verification to qualitative procedures  need move towards use of objective scores!! There are no formal WMO verification procedures, but some guidance on procedures is being published by WMO CCl Focus on how well forecasts correspond with observations (quality), and also on attributes making forecasts potentially useful (value). Small sample sizes (few years, few stations) typical of seasonal forecasts  large sampling errors

What is a good forecast? (Murphy 1993) 3 types of goodness: CONSISTENCY  true indication of what the forecaster thinks is going to happen QUALITY  how well what was forecast corresponds with what happened VALUE/UTILITY  “value” economic, or social, or otherwise.

Probabilistic forecasts and forecast quality A forecaster says there is a 100% chance of rain tomorrow  It rains  Very good forecast! A forecaster says there is a 80% chance of rain tomorrow  It rains  ? A forecaster says there is a 50% chance of rain tomorrow  It rains  ? A forecaster says there is a 10% chance of rain tomorrow  It rains  ? How good are the different forecast?

One reasonably common practice is to define probabilistic forecasts as “correct” if the category with the highest probability verified. Most RCOFs verify qualitatively in this way Forecasters typically become tempted to hedge towards issuing higher probabilities on the normal category to avoid a two category “error”  Scoring strategy is an issue!!

Verification procedures suitable for the forecasts in the format in which they are presented. If forecasts are delivered in form of tercile- based categories  Verification should fit to it!

Attributes of “good” probabilistic forecasts (Murphy 1993) Resolution Does the outcome change when the forecast changes? OUTCOME CONDITIONED BY FORECAST Example: does above-normal rainfall become more frequent when its probability increases? Discrimination Does the forecast differ when the outcome differs? FORECAST CONDITIONED BY OUTCOME Example: is the probability on above-normal rainfall higher when above-normal rainfall occurs? Reliability if observation falls in the category as FREQUENTLY as the forecast implies Sharpness Probabilities differing MARKEDLY from the climatology Skill It COMPARES two forecasts with some metric

From EUMETCAL( ) High reliabilityHigh resolutionHigh SharpnessDiscriminatoryHigh Skill The forecaster predicts the long term climatological frequency on each occasion The forecaster predicts categorically, that is, he assigns a forecast of 100% to the category he thinks is most likely, and 0 to the other. The forecaster manages to forecast 45% probability when the event does not occur and 55% when it does. A forecaster who is sure, but never absolutely certain, forecasting 80% when he thinks rain will occur and 20% when he thinks it won’t. The forecaster sits back with a smile on his face: He went out on a limb and predicted 90% probability of rain in his dry climate where it normally rains on only 10% of the days. And sure enough, it rained.

Recommended scores/procedures for series of forecasts (*) Minimum set for an operational centre

ROC curves: idealized examples (a) good discrimination and good skill (c) excellent discrimination (e) no discrimination (g) good discrimination for low probability forecasts (b) good discrimination but bad skill (d) good discrimination (f) good discrimination for high probability forecasts (h) good discrimination for confident (high and low probability) forecasts.

Simple realistic example

Reliability diagrams: observed relative freq. vs forecasted relative freq. (a) perfect reliability, (c) over-confidence (e) under-forecasting (b) no resolution (d) under-confidence (f) overforecasting

Reliability diagrams for the first 10 years of PRESAO (seasonal rainfall forecasts Jul-Sept) Over-forecasting in normal category Under-forecasting in below normal category Weak sharpness

Verification with CPT

Verification of tercile-based forecasts only requires information of the obs. category  problems related data policy circumvected

Scores to verify the consensus forecasts and scores to improve the consensus process

Reference climatology is relevant! Paco’s tranparency!! Tercile-based seasonal forecasts refered to a climatology Climatologist  long reference periods (30 y) Users  short (10 y) recient periods

Recommendations Assess the degree to which forecasts are being hedged on normal  Eliminate, or at least reduce, the hedging: – Use “proper” scoring procedures – Review procedures for setting probabilities Agree upon a minimum set of verification procedures for RCOF products. Encourage greater standardization in forecast production

Proposal Start with a minimum verification package (following WMO-CCI guidelines) verifying consensus forecast (tercile-based) produced so far by SEECOF and PRESANORD Use initially ECA&D data from a set of selected stations and tercile-based obs. (A, N, B) Agree on a reference period to establish our tercile values Report on MedCOF-2

THANK YOU FOR YOUR ATTENTION! and discussion on RCOF verification to be continued!!!

Discrimination Perfect Very bad % T % T % F % T % F % F % T % T % F % F % F % T % T % T % F % T % T % F % F % T If prob>35% always T