1 James Brown An introduction to verifying probability forecasts RFC Verification Workshop.

Slides:



Advertisements
Similar presentations
Medium-range Ensemble Streamflow forecast over France F. Rousset-Regimbeau (1), J. Noilhan (2), G. Thirel (2), E. Martin (2) and F. Habets (3) 1 : Direction.
Advertisements

Model Evaluation Tools MET. What is MET Model Evaluation Tools ( MET )- a powerful and highly configurable verification package developed by DTC offering:
Bill Lawrence – DOH. “Hydrologic Model Output Statistics”  Current short term ensembles have proven unreliable, mainly because no hydrologic uncertainty.
1 Verification Continued… Holly C. Hartmann Department of Hydrology and Water Resources University of Arizona RFC Verification Workshop,
SIPR Dundee. © Crown copyright Scottish Flood Forecasting Service Pete Buchanan – Met Office Richard Maxey – SEPA SIPR, Dundee, 21 June 2011.
Key Considerations for Report Generation & Customization Richard Wzorek Director, Production IT Confidential © Almac Group 2012.
1 of Introduction to Forecasts and Verification.
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Experimental Real-time Seasonal Hydrologic Forecasting Andrew Wood Dennis Lettenmaier University of Washington Arun Kumar NCEP/EMC/CMB presented: JISAO.
Application of Forecast Verification Science to Operational River Forecasting in the National Weather Service Julie Demargne, James Brown, Yuqiong Liu.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Exploring the Use of Object- Oriented Verification at the Hydrometeorological Prediction Center Faye E. Barthold 1,2, Keith F. Brill 1, and David R. Novak.
EE694v-Verification-Lect5-1- Lecture 5 - Verification Tools Automation improves the efficiency and reliability of the verification process Some tools,
MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between.
Alan F. Hamlet Andy Wood Seethu Babu Marketa McGuire Dennis P. Lettenmaier JISAO Climate Impacts Group and the Department of Civil Engineering University.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
June 23, 2011 Kevin Werner NWS Colorado Basin River Forecast Center 1 NOAA / CBRFC Water forecasts and data in support of western water management.
Colorado Basin River Forecast Center Water Supply Forecasting Method Michelle Stokes Hydrologist in Charge Colorado Basin River Forecast Center April 28,
Evaluation of Potential Performance Measures for the Advanced Hydrologic Prediction Service Gary A. Wick NOAA Environmental Technology Laboratory On Rotational.
Water Supply Forecast using the Ensemble Streamflow Prediction Model Kevin Berghoff, Senior Hydrologist Northwest River Forecast Center Portland, OR.
Probabilistic forecasts of (severe) thunderstorms for the purpose of issuing a weather alarm Maurice Schmeits, Kees Kok, Daan Vogelezang and Rudolf van.
(date, event info here?)‏ SHARP Mark Love / Hydrologist / SERFC Supplemental Hydro- Analysis & Response Program July2007.
Copyright 2012, University Corporation for Atmospheric Research, all rights reserved Verifying Ensembles & Probability Fcsts with MET Ensemble Stat Tool.
3- System modelling An architectural model presents an abstract view of the sub-systems making up a system May include major information flows between.
ECMWF WWRP/WMO Workshop on QPF Verification - Prague, May 2001 NWP precipitation forecasts: Validation and Value Deterministic Forecasts Probabilities.
April 24, 2007 Nihat Cubukcu Utilization of Numerical Weather Forecast in Energy Sector.
Creating a Shared Vision Model. What is a Shared Vision Model? A “Shared Vision” model is a collective view of a water resources system developed by managers.
ESET ALEMU WEST Consultants, Inc. Bellevue, Washington.
Verification Summit AMB verification: rapid feedback to guide model development decisions Patrick Hofmann, Bill Moninger, Steve Weygandt, Curtis Alexander,
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
Forecasting Streamflow with the UW Hydrometeorological Forecast System Ed Maurer Department of Atmospheric Sciences, University of Washington Pacific Northwest.
Unit-5 Automated Comparison. VERIFICATION Verification and Validation are independent procedures that are used together for checking that a product, service,
Where the Research Meets the Road: Climate Science, Uncertainties, and Knowledge Gaps First National Expert and Stakeholder Workshop on Water Infrastructure.
National Weather Service Application of CFS Forecasts in NWS Hydrologic Ensemble Prediction John Schaake Office of Hydrologic Development NOAA National.
Requirements from KENDA on the verification NetCDF feedback files: -produced by analysis system (LETKF) and ‘stat’ utility ((to.
Variational Assimilation (VAR) Presented by: Jerry Nunn Hydrologist In Charge West Gulf River Forecast Center October 28, 2003.
Modeling Development CRFS—Technical Meeting November 14, 2012.
Use of Mesoscale Ensemble Weather Predictions to Improve Short-Term Precipitation and Hydrological Forecasts Michael Erickson 1, Brian A. Colle 1, Jeffrey.
CBRFC Stakeholder Forum February 24, 2014 Ashley Nielson Kevin Werner NWS Colorado Basin River Forecast Center 1 CBRFC Forecast Verification.
Logistical Verification Forecast Services in IHFS Mary Mullusky RFC Verification Workshop, August
Colorado Basin River Forecast Center and Drought Related Forecasts Kevin Werner.
Standard Verification Strategies Proposal from NWS Verification Team NWS Verification Team Draft03/23/2009 These slides include notes, which can be expanded.
RFC Climate Requirements 2 nd NOAA Climate NWS Dialogue Meeting January 4, 2006 Kevin Werner.
1 Probabilistic Forecast Verification Allen Bradley IIHR Hydroscience & Engineering The University of Iowa RFC Verification Workshop 16 August 2007 Salt.
Sources of Skill and Error in Long Range Columbia River Streamflow Forecasts: A Comparison of the Role of Hydrologic State Variables and Winter Climate.
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Alan F. Hamlet Andy Wood Dennis P. Lettenmaier JISAO Center for Science in the Earth System Climate Impacts Group and the Department.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
DOWNSCALING GLOBAL MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR FLOOD PREDICTION Nathalie Voisin, Andy W. Wood, Dennis P. Lettenmaier University of Washington,
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
EVALUATION OF A GLOBAL PREDICTION SYSTEM: THE MISSISSIPPI RIVER BASIN AS A TEST CASE Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier Civil and.
CHPS-XEFS Ensemble Pre- Processing (EPP) Update Prepared by DJ Seo Feb 23,
Long-lead streamflow forecasts: 2. An approach based on ensemble climate forecasts Andrew W. Wood, Dennis P. Lettenmaier, Alan.F. Hamlet University of.
Verification of C&V Forecasts Jennifer Mahoney and Barbara Brown 19 April 2001.
Overview of CBRFC Flood Operations Arizona WFOs – May 19, 2011 Kevin Werner, SCH.
Ensemble Forecasts Andy Wood CBRFC. Forecast Uncertainties Meteorological Inputs: Meteorological Inputs: Precipitation & temperature Precipitation & temperature.
RFC Verification Workshop What to Expect Julie Demargne and Mary Mullusky RFC Verification Workshop, August
National Oceanic and Atmospheric Administration’s National Weather Service Colorado Basin River Forecast Center Salt Lake City, Utah 11 The Hydrologic.
HIC Meeting, 02/25/2010 NWS Hydrologic Forecast Verification Team: Status and Discussion Julie Demargne OHD/HSMB Hydrologic Ensemble Prediction (HEP) group.
Long-Range Streamflow Forecasting Products and Water Resources Management Applications in the Columbia River Basin Alan F. Hamlet, Andy Wood, Dennis P.
2016 HEPEX Workshop Université Laval, Quebec, Canada
Verifying and interpreting ensemble products
Eric Jones Senior Hydrologist Lower Mississippi River Forecast Center
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Overview of Models & Modeling Concepts
Probabilistic forecasts
Application of a global probabilistic hydrologic forecast system to the Ohio River Basin Nathalie Voisin1, Florian Pappenberger2, Dennis Lettenmaier1,
N. Voisin, J.C. Schaake and D.P. Lettenmaier
Verification of Tropical Cyclone Forecasts
Presentation transcript:

1 James Brown An introduction to verifying probability forecasts RFC Verification Workshop

2 1.Introduction to methods What methods are available? How do they reveal (or not) particular errors? Lecture now, and hands-on training later 2.Introduction to prototype software Ensemble Verification System (EVS) Part of a larger experimental project (XEFS) Lecture now, and hands on training later Goals for today

3 3.To establish user-requirements EVS in very early (prototype) stage Pool of methods may expand or contract Need some input on verification products AND to address pre-workshop questions…... Goals for today

4 How is ensemble verification done? Same for short/long-term ensembles? What tools, and are they operational? Which metrics for which situations? Simple metrics for end-users? How best to manage the workload? What data need to be archived/how? Pre-workshop questions

5 1.Background and status 2.Overview of EVS 3.Metrics available in EVS 4.First look at the user-interface (GUI) Contents for next hour

6 1. Background and status

7 A first look at operational needs Two classes of verification identified 1.High time sensitivity (‘prognostic’) e.g. how reliable is my live flood forecast?... …where should I hedge my bets? 2.Less time sensitive (‘diagnostic’) e.g. which forecasts do less well and why? A verification strategy?

8 Prognostic example Temperature ( o C) Forecast lead day Live forecast (L) Historical observations | μ H = μ L ± 1.0 ˚ C Matching historical forecasts (H)

9 Diagnostic example Probability of warning correctly (hit) Probability of warning incorrectly (‘false alarms’) e.g. flood warning when P>=0.9 Climatology Single-valued forecast

10 Motivation for EVS (and XEFS) Demand: forecasters and their customers Demand for useable verification products ….limitations of existing software History Ensemble Verification Program (EVP) Comprised (too) many parts, lacked flexibility Prototype EVS begun in May 07 for XEFS….. Motivation for EVS

11 Position in XEFS IFP Ensemble Viewer OFS Raw flow ens. Pp’ed flow ens. Ensemble Verification Subsystem Flow Data Ens. Product Generation Subsystem Ensemble verification products Hydrologic Ensemble Hindcaster Ens. User Interface EPP User Interface Ens. Pre- Processor Atmospheric forcing data Ensemble /prob. products Ens. Post- Proc. Ens. Streamflow Prediction Subsystem HMOS Ensemble Processor MODs EPP3 ESP2EnsPostEPG EVS Hydro- meteorol. ensembles Precip., temp. etc. Streamflow

12 2. Overview of EVS

13 Diagnostic verification For diagnostic purposes (less time-sensitive) Prognostic built into forecasting systems Diagnostic questions include…. Are ensembles reliable? Prob[flood]=0.9: does it occur 9/10 times? Are forecaster MODS working well? What are the major sources of uncertainty? Scope of EVS

14 Verification of continuous time-series Temperature, precipitation, streamflow etc. > 1 forecast points, but not spatial products All types of forecast times Any lead time (e.g. 1 day – 2 years or longer) Any forecast resolution (e.g. hourly, daily) Pair forecasts/observed (in different t-zones) Ability to aggregate across forecast points Design goals of EVS

15 Flexibility to target data of interest Subset based on forecasts and observations Two conditions: 1) time; 2) variable value e.g. forecasts where ensemble mean < 0˚C e.g. max. observed flow in 90 day window Ability to pool/aggregate forecast points Number of observations can be limiting Sometimes appropriate to pool points Design goals of EVS

16 Carefully selected metrics Different levels of detail on errors Some are more complex than others, but…. Use cases and online docs. to assist To be ‘user-friendly’ Many factors determine this…. GUI, I/O, exec. speed, batch modes Design goals of EVS

17 Example of workflow (Re)define verification unit Condition: Flow > flood && winter Box plot and Talagrand Run (data read, paired, plots done) View and save plots How biased are my winter flows > flood level at dam A?

18 Coordinated across XEFS: The forecasts Streamflow: ESP binary files (.CS) Temperature and precip: OHD datacard files The observations OHD datacard files Unlikely to be database in near future Archiving requirements

19 3. Metrics available

20 Many ways to test a probability forecast 1.Tests for single-valued property (e.g. mean) 2.Tests of broader forecast distribution Both may involve reference forecasts (“skill”) Caveats in testing probabilities Observed probabilities require many events Big assumption 1: we can ‘pool’ events Big assumption 2: observations are ‘good’ Types of metrics

21 Discrete/categorical forecasts Many metrics rely on discrete forecasts e.g. will it rain? {yes/no} (rain > 0.01) e.g. will it flood? {yes/no} (stage > flood level) What about continuous forecasts? An infinite number of events Arbitrary event thresholds (i.e. ‘bins’)? Typically, yes (and choice will affect results) Problem of cont. forecasts

22 Detail varies with verification question e.g. inspection of ‘blown’ forecasts (detailed) e.g. avg. reliability of flood forecast (< detail) e.g. rapid screening of forecasts (<< detail) All included to some degree in EVS…… Metrics in EVS

23 Greatest + ve 90 percent. 80 percent. 50 percent. 20 percent. 10 percent. ‘Errors’ for 1 forecast Greatest - ve Observation Ensemble forecast errors Most detailed (box plot) Time (days since start time)

24 Greatest + ve 90 percent. 80 percent. 50 percent. 20 percent. 10 percent. ‘Errors’ for 1 forecast Greatest - ve Observation Ensemble forecast errors Observed value (increasing size) Most detailed (box plot)

25 Less detail (Reliability) Observed probability given forecast Forecast probability (probability of flooding) “On occasions when flooding is forecast with probability 0.5, it should occur 50% of the time.” “Forecast bias”

26 Less detail (C. Talagrand) Cumulative probability “If river stage <=X is forecast with probability 0.5, it should be observed 50% of the time.” Position of observation in forecast distribution “Forecast bias”

27 Least detailed (a score) River stage Time (days) Flood stage Forecast Observation Brier score = 1/5 x { ( ) 2 + ( ) 2 + ( ) 2 + ( ) 2 + ( ) 2 } 4

28 Least detailed (a score) Cumulative probability Precipitation amount Single forecast Observation A B CRPS = A 2 + B 2 Then average across multiple forecasts: small scores are better

29 4. First look at the GUI

30 Two-hour lab sessions with EVS Start with synthetic data (with simple errors) Then more on to a couple of real cases Verification plans and feedback Real-time (‘prognostic’) verification Screening verification outputs Developments in EVS Feedback: discussion and survey Rest of today