© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Ensemble-4DWX.

Slides:



Advertisements
Similar presentations
Climate Modeling LaboratoryMEASNC State University An Extended Procedure for Implementing the Relative Operating Characteristic Graphical Method Robert.
Advertisements

ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF.
Slide 1ECMWF forecast products users meeting – Reading, June 2005 Verification of weather parameters Anna Ghelli, ECMWF.
Robin Hogan Ewan OConnor University of Reading, UK What is the half-life of a cloud forecast?
Model Evaluation Tools MET. What is MET Model Evaluation Tools ( MET )- a powerful and highly configurable verification package developed by DTC offering:
Comparing and Contrasting Post-processing Approaches to Calibrating Ensemble Wind and Temperature Forecasts Tom Hopson Luca Delle Monache, Yubao Liu, Gregory.
Verification of probability and ensemble forecasts
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Daria Kluver Independent Study From Statistical Methods in the Atmospheric Sciences By Daniel Wilks.
Gridded OCF Probabilistic Forecasting For Australia For more information please contact © Commonwealth of Australia 2011 Shaun Cooper.
Instituting Reforecasting at NCEP/EMC Tom Hamill (ESRL) Yuejian Zhu (EMC) Tom Workoff (WPC) Kathryn Gilbert (MDL) Mike Charles (CPC) Hank Herr (OHD) Trevor.
1 Operational low visibility statistical prediction Frédéric Atger (Météo-France)
Recent performance statistics for AMPS real-time forecasts Kevin W. Manning – National Center for Atmospheric Research NCAR Earth System Laboratory Mesoscale.
Barbara Casati June 2009 FMI Verification of continuous predictands
¿How sensitive are probabilistic precipitation forecasts to the choice of the ensemble generation method and the calibration algorithm? Juan J. Ruiz 1,2.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
Evaluation of Potential Performance Measures for the Advanced Hydrologic Prediction Service Gary A. Wick NOAA Environmental Technology Laboratory On Rotational.
Performance of the MOGREPS Regional Ensemble
Copyright 2012, University Corporation for Atmospheric Research, all rights reserved Verifying Ensembles & Probability Fcsts with MET Ensemble Stat Tool.
Verification of ensembles Courtesy of Barbara Brown Acknowledgments: Tom Hamill, Laurence Wilson, Tressa Fowler Copyright UCAR 2012, all rights reserved.
Rank Histograms – measuring the reliability of an ensemble forecast You cannot verify an ensemble forecast with a single.
Operational Flood Forecasting for Bangladesh: Tom Hopson, NCAR Peter Webster, GT A. R. Subbiah and R. Selvaraju, ADPC Climate Forecast Applications for.
Tutorial. Other post-processing approaches … 1) Bayesian Model Averaging (BMA) – Raftery et al (1997) 2) Analogue approaches – Hopson and Webster, J.
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
Verification methods - towards a user oriented verification WG5.
Improving Ensemble QPF in NMC Dr. Dai Kan National Meteorological Center of China (NMC) International Training Course for Weather Forecasters 11/1, 2012,
Celeste Saulo and Juan Ruiz CIMA (CONICET/UBA) – DCAO (FCEN –UBA)
Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
Latest results in verification over Poland Katarzyna Starosta, Joanna Linkowska Institute of Meteorology and Water Management, Warsaw 9th COSMO General.
Requirements from KENDA on the verification NetCDF feedback files: -produced by analysis system (LETKF) and ‘stat’ utility ((to.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Accounting for Change: Local wind forecasts from the high-
Multi-Model or Post- processing: Pros and Cons Tom Hopson - NCAR Martyn Clark - NIWA Andrew Slater - CIRES/NSIDC.
61 st IHC, New Orleans, LA Verification of the Monte Carlo Tropical Cyclone Wind Speed Probabilities: A Joint Hurricane Testbed Project Update John A.
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
Short-Range Ensemble Prediction System at INM José A. García-Moya SMNT – INM 27th EWGLAM & 12th SRNWP Meetings Ljubljana, October 2005.
. Outline  Evaluation of different model-error schemes in the WRF mesoscale ensemble: stochastic, multi-physics and combinations thereof  Where is.
Refinement and Evaluation of Automated High-Resolution Ensemble-Based Hazard Detection Guidance Tools for Transition to NWS Operations Kick off JNTP project.
Quantile regression as a means of calibrating and verifying a mesoscale NWP ensemble Tom Hopson 1 Josh Hacker 1, Yubao Liu 1, Gregory Roux 1, Wanli Wu.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
1 NUOPC National Unified Operational Prediction Capability Update to COPC 4 – 5 Nov 2015 Dave McCarren, NUOPC DPM.
Statistical Post Processing - Using Reforecast to Improve GEFS Forecast Yuejian Zhu Hong Guan and Bo Cui ECM/NCEP/NWS Dec. 3 rd 2013 Acknowledgements:
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss A more reliable COSMO-LEPS F. Fundel, A. Walser, M. A.
Classification Ensemble Methods 1
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Nathalie Voisin 1, Florian Pappenberger 2, Dennis Lettenmaier 1, Roberto Buizza 2, and John Schaake 3 1 University of Washington 2 ECMWF 3 National Weather.
Common verification methods for ensemble forecasts
Verification of ensemble precipitation forecasts using the TIGGE dataset Laurence J. Wilson Environment Canada Anna Ghelli ECMWF GIFS-TIGGE Meeting, Feb.
Diagnostic verification and extremes: 1 st Breakout Discussed the need for toolkit to build beyond current capabilities (e.g., NCEP) Identified (and began.
Extracting probabilistic severe weather guidance from convection-allowing model forecasts Ryan Sobash 4 December 2009 Convection/NWP Seminar Series Ryan.
DOWNSCALING GLOBAL MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR FLOOD PREDICTION Nathalie Voisin, Andy W. Wood, Dennis P. Lettenmaier University of Washington,
VERIFICATION OF A DOWNSCALING SEQUENCE APPLIED TO MEDIUM RANGE METEOROLOGICAL PREDICTIONS FOR GLOBAL FLOOD PREDICTION Nathalie Voisin, Andy W. Wood and.
Verification methods - towards a user oriented verification The verification group.
Google Meningitis Modeling Tom Hopson October , 2010.
LEPS VERIFICATION ON MAP CASES
Tom Hopson, NCAR (among others) Satya Priya, World Bank
Ensemble Forecasting: Calibration, Verification, and use in Applications Tom Hopson.
Verifying and interpreting ensemble products
Tom Hopson, Jason Knievel, Yubao Liu, Gregory Roux, Wanli Wu
Precipitation Products Statistical Techniques
Binary Forecasts and Observations
Winter storm forecast at 1-12 h range
Nathalie Voisin, Andy W. Wood and Dennis P. Lettenmaier
Google Meningitis Modeling
Probabilistic forecasts
COSMO-LEPS Verification
Ensemble-4DWX update: focus on calibration and verification
Can we distinguish wet years from dry years?
Rapid Adjustment of Forecast Trajectories: Improving short-term forecast skill through statistical post-processing Nina Schuhen, Thordis L. Thorarinsdottir.
Short Range Ensemble Prediction System Verification over Greece
Presentation transcript:

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Ensemble-4DWX update: focus on calibration and verification

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Summary of progress Upgraded from WRF V2.2.1 to V Replaced 3 physics members (slab LSM, Grell-Dvevenyi Cu, and Betts-Miller Cu schemes); added RUC LSM, no horizontal diffusion, Thomson scheme with positive-definite advection Configured E-4DWX for ATC and operated for a 2 month period (Dec 1 and Jan 30) Configured and run E-4DWX for supporting UT Dept of Air Quality Improved MYJ and YSU PBL height diagnosis and PBL mixing Added new graphics and improve post-processing flexibility (installation for GMOD and plotting historical case archive) and computing parallelisms Presented the system at a number of AMS and other conferences Continued R&D of on-line verification and calibration

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Ongoing work and plans E-4DWX science paper for MWR; technology brief for BAMS Member performance evaluation for 2009 Feb and Mar, and compare to the results we obtained for 2008 Feb and Mar; look for improvements Member-based evaluation of ATC E-4DWX run to identify best WRF model configuration ATC cold-air damming study using E-4DWX archive Begin development of 4d-ENKF

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory New product examples

© 2009 UCAR. All rights reserved. New product examples ATEC-4DWX IPR, 21−22 April National Security Applications Program Research Applications Laboratory

© 2009 UCAR. All rights reserved. New product examples ATEC-4DWX IPR, 21−22 April National Security Applications Program Research Applications Laboratory

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Calibration and verification Ensemble calibration to correct predicted distribution. Calibration is needed for users capable of decision making with probabilistic guidance. Will be needed for foreseeable future. Verification of different ensemble characteristics is easily completed when performing calibration.

NCAR/RAL - National Security Applications Program 8 What do we mean by “calibration” or “post-processing”? Probability calibration Temperature [K] Probability Temperature [K] Post-processing has corrected: the “on average” bias as well as under-representation of the 2nd moment of the empirical forecast PDF (i.e. corrected its “dispersion” or “spread”) “spread” or “dispersion” “bias” obs Forecast PDF Forecast PDF

NCAR/RAL - National Security Applications Program 9 Benefits of Post-Processing Essential for tailoring to local application:  NWP provides spatially- and temporally-averaged gridded forecast output => Applying gridded forecasts to point locations requires location specific calibration to account for spatial- and temporal- variability ( => increasing ensemble dispersion) => Relatively inexpensive!

NCAR/RAL - National Security Applications Program 10 Example of Quantile Regression (QR) Our application Fitting T quantiles using QR conditioned on: 1)Reforecast ens 2)ensemble mean 3)ensemble median 4) ensemble stdev 5) Persistence 6) Log Reg quantile

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Summary of progress Calibration (and Verification) Scheme fully-automated Calibrating Temperature and Dew Point Temperature Calibration specific for unique set of models that are available for each cycle Utilizes “persistence” if available 26 sites over DPG Full calibration for all sites ~ 10hrs for each weather variable Using lookup tables ~ 1hr Updating tables once per week

NCAR/RAL - National Security Applications Program 12 Calibration Procedure 1)Fit Logistic Regression (LR) ensembles –Calibrate CDF over prescribed set of climatological quantiles –For each forecast: resample 15 member ensemble set For each quantile: 2)Perform a “climatological” fit to the data 3)Starting with full regressor set, iteratively select best subset using “step-wise cross-validation” –Fitting done using QR –Selection done by: a)Minimizing QR cost function b)Satisfying the binomial distribution ( 2nd pass: segregate forecasts into differing ranges of ensemble dispersion, and refit models ) Probability Temperature [K]  obs Forecast PDF T [K] Time Forecastsobserved Regressors for each quantile: 1) reforecast ensemble 2) ens mean 3) ens median 4) ens stdev 5) persistence 6) logistic regression quantile

NCAR/RAL - National Security Applications Program 13 Verifying ensemble forecasts Measures Used in automated model selection: 1)Rank histogram 2)Root Mean square error (RMSE) 3)Brier score 4)Rank Probability Score (RPS) 5)Relative Operating Characteristic (ROC) curve

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory 6hr Temperature Time-series Before Calibration After Calibration

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory 36hr Temperature Time-series Before Calibration After Calibration

NCAR/RAL - National Security Applications Program 16 Raw versus Calibrated PDF’s obs Blue is “raw” ensemble Black is calibrated ensemble Red is the observed value Notice: significant change in both “bias” and dispersion of final PDF (also notice PDF asymmetries)

Troubled Rank Histograms Slide from Matt Pocernic Ensemble # Ensemble # Counts Counts

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory 6hr Temperature Rank Histograms

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory 36hr Temperature Rank Histograms

NCAR/RAL - National Security Applications Program 20 Verifying ensemble forecasts Measures Used: 1)Rank histogram 2)Root Mean square error (RMSE) 3)Brier score 4)Rank Probability Score (RPS) 5)Relative Operating Characteristic (ROC) curve => Using these for automated calibration model selection

NCAR/RAL - National Security Applications Program 21 Rank Probability Score for multi-categorical or continuous variables

Skill Scores Single value to summarize performance. Reference forecast - best naive guess; persistence, climatology A perfect forecast implies that the object can be perfectly observed Positively oriented – Positive is good

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory 36hr Temperature Time-series RMSE Skill Score CRPS Skill Score Reference Forecasts: Black -- raw ensemble Blue -- persistence

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory RMSE of Models 6hr Lead-time 36hr Lead-time

© 2009 UCAR. All rights reserved. ATEC-4DWX IPR, 21−22 April 2009 National Security Applications Program Research Applications Laboratory Significant Calibration Regressors 6hr Lead-time 36hr Lead-time

NCAR/RAL - National Security Applications Program 26 Future Plans  Correct over-dispersion of calibration  Implement procedure for a) wind speed, b) wind direction, c) precipitation, d) pressure  Diagnose most informative model set to use operationally  Develop Scheme for model points without surface observations => over whole model-gridded domain

Continuous scores: MSE Average of the squares of the errors: it measures the magnitude of the error, weighted on the squares of the errors it does not indicate the direction of the error Quadratic rule, therefore large weight on large errors:  good if you wish to penalize large error  sensitive to large values (e.g. precipitation) and outliers; sensitive to large variance (high resolution models); encourage conservative forecasts (e.g. climatology) Attribute: measures accuracy Slide from Barbara Casati => For ensemble forecast, use ensemble mean

Scatter-plot and Contingency Table Does the forecast detect correctly temperatures above 18 degrees ? Slide from Barbara Casati Brier Score y = forecasted event occurence o = observed occurrence (0 or 1) i = sample # of total n samples => Note similarity to MSE

Conditional Distributions Conditional histogram and conditional box-plot Slide from Barbara Casati

Scatter-plot and Contingency Table Does the forecast detect correctly temperatures above 18 degrees ? Does the forecast detect correctly temperatures below 10 degrees ? Slide from Barbara Casati

Discrimination Plot Outcome = No False Alarms Outcome = Yes Hits Decision Threshold Slide from Matt Pocernic

Receiver Operating Characteristic (ROC) Curve