Verification methods - towards a user oriented verification The verification group.

Slides:



Advertisements
Similar presentations
Robin Hogan Ewan OConnor University of Reading, UK What is the half-life of a cloud forecast?
Advertisements

Chapter 3 Properties of Random Variables
VERIFICATION Highligths by WG5. 9° General MeetingAthens September Working package/Task on “standardization” The “core” Continuous parameters: T2m,
6th WMO tutorial Verification Martin GöberContinuous 1 Good afternoon! नमस्कार नमस्कार Guten Tag! Buenos dias! до́брый день! до́брыйдень Qwertzuiop asdfghjkl!
14 May 2001QPF Verification Workshop Verification of Probability Forecasts at Points WMO QPF Verification Workshop Prague, Czech Republic May 2001.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Becoming Acquainted With Statistical Concepts CHAPTER CHAPTER 12.
Review of Basics. REVIEW OF BASICS PART I Measurement Descriptive Statistics Frequency Distributions.
Review of Basics. REVIEW OF BASICS PART I Measurement Descriptive Statistics Frequency Distributions.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Verification of DWD Ulrich Damrath & Ulrich Pflüger.
Reliability Trends of the Global Forecast System Model Output Statistical Guidance in the Northeastern U.S. A Statistical Analysis with Operational Forecasting.
A quick introduction to the analysis of questionnaire data John Richardson.
MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between.
Barbara Casati June 2009 FMI Verification of continuous predictands
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Relationships Among Variables
COSMO General Meeting Zurich, 2005 Institute of Meteorology and Water Management Warsaw, Poland- 1 - Verification of the LM at IMGW Katarzyna Starosta,
COSMO General Meeting – Moscow Sept 2010 Some results from operational verification in Italy Angela Celozzi - Federico Grazzini Massimo Milelli -
Chapter 2: The Research Enterprise in Psychology
Verification of ensembles Courtesy of Barbara Brown Acknowledgments: Tom Hamill, Laurence Wilson, Tressa Fowler Copyright UCAR 2012, all rights reserved.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Eidgenössisches Departement des Innern EDI Bundesamt für Meteorologie und Klimatologie MeteoSchweiz Statistical Characteristics of High- Resolution COSMO.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 10. Hypothesis Testing II: Single-Sample Hypothesis Tests: Establishing the Representativeness.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
Verification methods - towards a user oriented verification WG5.
SEASONAL COMMON PLOT SCORES A DRIANO R ASPANTI P ERFORMANCE DIAGRAM BY M.S T ESINI Sibiu - Cosmo General Meeting 2-5 September 2013.
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
Latest results in verification over Poland Katarzyna Starosta, Joanna Linkowska Institute of Meteorology and Water Management, Warsaw 9th COSMO General.
The latest results of verification over Poland Katarzyna Starosta Joanna Linkowska COSMO General Meeting, Cracow September 2008 Institute of Meteorology.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Priority project « Advanced interpretation and verification.
Objectives 2.1Scatterplots  Scatterplots  Explanatory and response variables  Interpreting scatterplots  Outliers Adapted from authors’ slides © 2012.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Accounting for Change: Local wind forecasts from the high-
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Ping Zhu, AHC5 234, Office Hours: M/W/F 10AM - 12 PM, or by appointment M/W/F,
Probabilistic Forecasting. pdfs and Histograms Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density,
. Outline  Evaluation of different model-error schemes in the WRF mesoscale ensemble: stochastic, multi-physics and combinations thereof  Where is.
Deutscher Wetterdienst Bootstrapping – using different methods to estimate statistical differences between model errors Ulrich Damrath COSMO GM Rome 2011.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss Local Probabilistic Weather Predictions for Switzerland.
Introduction to Basic Statistical Tools for Research OCED 5443 Interpreting Research in OCED Dr. Ausburn OCED 5443 Interpreting Research in OCED Dr. Ausburn.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss A more reliable COSMO-LEPS F. Fundel, A. Walser, M. A.
Introducing Communication Research 2e © 2014 SAGE Publications Chapter Seven Generalizing From Research Results: Inferential Statistics.
U. Damrath, COSMO GM, Athens 2007 Verification of numerical QPF in DWD using radar data - and some traditional verification results for surface weather.
Verification of ensemble systems Chiara Marsigli ARPA-SIMC.
Quality Control: Analysis Of Data Pawan Angra MS Division of Laboratory Systems Public Health Practice Program Office Centers for Disease Control and.
Overview of WG5 activities and Conditional Verification Project Adriano Raspanti - WG5 Bucharest, September 2006.
VERIFICATION Highligths by WG5. 2 Outlook Some focus on Temperature with common plots and Conditional Verification Some Fuzzy verification Long trends.
10th COSMO General Meeting, Cracow, Poland Verification of COSMOGR Over Greece 10 th COSMO General Meeting Cracow, Poland.
Latest results in the precipitation verification over Northern Italy Elena Oberto, Marco Turco, Paolo Bertolotto (*) ARPA Piemonte, Torino, Italy.
Chapter 6 Becoming Acquainted With Statistical Concepts.
LESSON 5 - STATISTICS & RESEARCH STATISTICS – USE OF MATH TO ORGANIZE, SUMMARIZE, AND INTERPRET DATA.
VERIFICATION Highligths by WG5. 2 Outlook The COSMO-Index COSI at DWD Time series of the index and its DWD 2003.
Forecasting. Model with indicator variables The choice of a forecasting technique depends on the components identified in the time series. The techniques.
Becoming Acquainted With Statistical Concepts
Current verification results for COSMO-EU and COSMO-DE at DWD
Verifying and interpreting ensemble products
Question 1 Given that the globe is warming, why does the DJF outlook favor below-average temperatures in the southeastern U. S.? Climate variability on.
Verification Overview
Christoph Gebhardt, Zied Ben Bouallègue, Michael Buchhold
15.1 The Role of Statistics in the Research Process
Some Verification Highlights and Issues in Precipitation Verification
Ulrich Pflüger & Ulrich Damrath
the performance of weather forecasts
Short Range Ensemble Prediction System Verification over Greece
VERIFICATION OF THE LAMI AT CNMCA
Presentation transcript:

Verification methods - towards a user oriented verification The verification group

COSMO GM Cracow September 2008 Course of verification on principal Verification Observation Forecast Data control Enduser Modeller Analysis Data control

COSMO GM Cracow September 2008 What are the results of verification? RMSE | S1 |ANOC ETS FBI FSS BIAS BSS ROC ISS

COSMO GM Cracow September 2008 What is a good forecast? A.H. Murphy, Weather and Forecasting, 8(1993), Iss.2,  Type 1 of goodness: Consistency  Correspondence between forecasts and judgements  Type 2 of goodness: Quality  Comparison of results for different seasons  Type 3 of goodness: Value  Incremental benefits of forecasts to users

COSMO GM Cracow September 2008 What is a good forecast? A.H. Murphy, Weather and Forecasting, 8(1993), Iss.2,  Type 1 of goodness: Consistency  Correspondence between forecasts and judgements  Type 2 of goodness: Quality  Comparison of results for different reasons  Type 3 of goodness: Value  Incremental benefits of forecasts to users

COSMO GM Cracow September 2008 Attributes of a forecasts related to observations (I)  BIAS  Confidence between mean forecast and mean observation (BIAS, FBI)  Association  Overall strength of linear relationships between individual pairs of observation and forecast (correlation)  Accuracy  Average correspondence between individual pairs of observation and forecast (mean absolute error, root mean square error, Brier score,...)  Skill  Accuracy of a forecast relative to a reference forecast (reduction of variance, Brier skill score)  Reliability  Correspondence between conditional mean observation and and conditioning forecast averaged over all forecasts (elements of conditional verification, probability forecasts)

COSMO GM Cracow September 2008 Attributes of a forecasts related to observations (II)  Resolution  Difference between conditional mean obseration and unconditional mean observation averaged over all forecasts (parts of the Brier score and the mean square error, probability forecasts)  Sharpness  Variability of forecasts as described by distribution of forecasts (parameter of forecast distribution: mean, standard deviation, skewness, kurtosis, median, quantiles,..., probability forecasts)  Discrimination I  Correspondence between conditional mean forecast and conditioning observation averaged over all observations (elements of conditional verification, probability forecasts)  Discrimination II  Difference between conditional mean forecast and unconditional mean forecast averaged over all obsercations (elements of conditional verification probability forecasts, )  Uncertainty  Variability of observations as described by distribution of observations (parameter of observation distribution: mean, standard deviation, skewness, kurtosis, median, quantiles,...)

COSMO GM Cracow September 2008 Attributes of a forecasts related to observations(I)  Bias - the correspondence between the mean forecast and mean observation.  Association - the strength of the linear relationship between the forecasts and observations (for example, the correlation coefficient measures this linear relationship)  Accuracy - the level of agreement between the forecast and the truth (as represented by observations). The difference between the forecast and the observation is the error. The lower the errors, the greater the accuracy.  Skill - the relative accuracy of the forecast over some reference forecast. The reference forecast is generally an unskilled forecast such as random chance, persistence (defined as the most recent set of observations, "persistence" implies no change in condition), or climatology. Skill refers to the increase in accuracy due purely to the "smarts" of the forecast system. Weather forecasts may be more accurate simply because the weather is easier to forecast -- skill takes this into account. referring to: A.H. Murphy, Weather and Forecasting, 8(1993), Iss.2,

COSMO GM Cracow September 2008  Reliability - the average agreement between the forecast values and the observed values. If all forecasts are considered together, then the overall reliability is the same as the bias. If the forecasts are stratified into different ranges or categories, then the reliability is the same as the conditional bias,  Resolution - the ability of the forecast to sort or resolve the set of events into subsets with different frequency distributions. This means that the distribution of outcomes when "A" was forecast is different from the distribution of outcomes when "B" is forecast. Even if the forecasts are wrong, the forecast system has resolution if it can successfully separate one type of outcome from another. Attributes of a forecasts related to observations(II)

COSMO GM Cracow September 2008  Sharpness - the tendency of the forecast to predict extreme values. To use a counter-example, a forecast of "climatology" has no sharpness. Sharpness is a property of the forecast only, and like resolution, a forecast can have this attribute even if it's wrong (in this case it would have poor reliability).  Discrimination - ability of the forecast to discriminate among observations, that is, to have a higher prediction frequency for an outcome whenever that outcome occurs.  Uncertainty - the variability of the observations. The greater the uncertainty, the more difficult the forecast will tend to be. Attributes of a forecasts related to observations(III)

COSMO GM Cracow September 2008 Current focal points of verification  Spatial verification methods  object oriented methods  „fuzzy“- techniques  Verification of probabilistic and ensemble forecasts  ensemble pdf  generic probability forecasts  probability of an event  Verification of extreme (rare) events  high-impact events  Operational verification  evaluation and monitoring  User-oriented verification strategies  tailored verification for any user  Forecast value  cost - loss analysis, development of an universal score  Verification packages  VERSUS, MET

COSMO GM Cracow September 2008 Some problems concerning significance  Most scores have a statistical outfit and seem to be open for significance tests.  Traditional significance tests require a defined number of degrees of freedom.  In most cases observations, forecasts and errors are correlated.  Therefore, the degrees of freedom cannot be obtained easily.  One way out: resampling and bootstrapping  What about statistical significance and meteorological significance?

COSMO GM Cracow September 2008 User-oriented verification strategies  What is the interest of an user?  Administrator:  Did forecasts yield to better results during last period of interest and in general?  What type of focal points for model development are of current interest? ...  Modeller  What type of errors occur in general?  What are the reasons for such errors?  How should the model modified in order to avoid or to reduce these errors?  If I have found the reason(s) for the error(s) and I have reduced their effect(s), is the forecast then improved? ...  External user  How can I interpret the forecasts?  What is the benefit of forecasts for me? ...

COSMO GM Cracow September 2008 User-oriented verification step by step 1. Diagnosis of errors normally done by examining the BIAS or the FBI The problem - mean values of observed and forecasted T2m over Germany during Sommer 2005 and Winter 2005/2006 (RMSE/STDV) Summer 2005 Winter 2005/2006

COSMO GM Cracow September 2008 User-oriented verification step by step 1. Diagnosis of errors normally done by examining the BIAS or the FBI The problem - mean values of observed and forecasted gusts over Germany during Spring 2007 (RMSE/STDV) COSMO-DE COSMO-EU

COSMO GM Cracow September 2008 User-oriented verification step by step 1. Diagnosis of errors normally done by examining the BIAS or the FBI Examples for four scores in four stylisized situations:

COSMO GM Cracow September 2008 An example for conditional verification Forecasted and observed values of surface level pressure over the region of Germany during DJF 2005/2006 (RMSE and STDV) Forecasted and observed values of surface level pressure over the region of Germany during DJF 2005/2006 observed and forecasted values lower than 1020 hPa (RMSE and STDV) Forecasted and observed values of surface level pressure over the region of Germany during DJF 2005/2006 observed and forecasted values higher than 1020 hPa (RMSE and STDV)

COSMO GM Cracow September 2008 User-oriented verification step by step 2. Some changes made by modellers  New diagnosis of gusts  to reduce the overestimation of gusts:  use wind at 10 m instead of interpolated wind from 30 m to compute gusts  New diagnosis of temperature 2m  to reduce the strong negative bias during winter and get a more realistic diurnal cycle:  set z 0 to 2 cm over land and change some other parameterizations  New SSO scheme (currently under examination)

COSMO GM Cracow September 2008 User-oriented verification step by step 3. The effects  New diagnosis of gusts  The overestimation of gusts is now reduced.  But: Extreme gusts are underestimated.  New diagnosis of temperature 2m  to reduce the strong negative bias during winter and get a more realistic diurnal cycle:  set z 0 to 2 cm over land and change some other parameterizations  New SSO scheme (currently under examination)

COSMO GM Cracow September 2008 User-oriented verification step by step 4. The proof of the effects: New diagnosis of gusts gusts > 12 ms -1 Böenverifikation der Experimente 6278 (COSMO-EU) Operational run 6301 (COSMO-DE) ETS FBI

COSMO GM Cracow September 2008 User-oriented verification step by step 4. The proof of the effects: New diagnosis of temperature 2m Comparison of COSMO-EU with experiment UTC: April/June 2007 : RMSE COSMO-EU area

COSMO GM Cracow September 2008 A basic law during model development: There are no gains without any losses! (maybe with some exceptions) Therefore, one has to look both at benefits and risks.

COSMO GM Cracow September 2008 One of known exceptions: The effect of a SSO scheme in COSMO-EU

COSMO GM Cracow September 2008 One of known exceptions: The effect of a SSO scheme in COSMO-EU

COSMO GM Cracow September 2008 User-oriented verification step by step 5. The risk: New diagnosis of gusts gusts > 25 ms -1 ETS FBI Böenverifikation der Experimente 6278 (COSMO-EU) Operational run 6301 (COSMO-DE)

COSMO GM Cracow September 2008 User-oriented verification step by step 5. The risk: New diagnosis of temperature 2m Comparison of COSMO-EU with experiment UTC: April/June 2007 : BIAS COSMO-EU area

COSMO GM Cracow September 2008 User-oriented verification step by step 6. The operational effect: New diagnosis of temperature 2m

COSMO GM Cracow September 2008 User-oriented verification step by step 7. The effect for administrators experessed in „The Score COSI“

COSMO GM Cracow September 2008 User-oriented verification step by step 7. The effect for administrators experessed in „The Score COSI“ asr Prognostic cloud ice Prognostic precipitation LME V 3.19 V 3.22 T 2m

COSMO GM Cracow September 2008 User-oriented verification step by step 8. Questions Are there any questions to me? Question from the verification group: What are the requirements to the verification process by users in order to make the process of model development as effective as possible?

COSMO GM Cracow September 2008 User-oriented verification step by step 7. The effect for administrators experessed in „The Score COSI“ asr

COSMO GM Cracow September 2008 User-oriented verification step by step 7. The effect for administrators experessed in „The Score COSI“