Presentation is loading. Please wait.

Presentation is loading. Please wait.

Verification and evaluation of a national probabilistic prediction system Barbara Brown NCAR 23 September 2009.

Similar presentations


Presentation on theme: "Verification and evaluation of a national probabilistic prediction system Barbara Brown NCAR 23 September 2009."— Presentation transcript:

1 Verification and evaluation of a national probabilistic prediction system Barbara Brown NCAR 23 September 2009

2 Goals… (Questions) What are desirable attributes of verification capabilities for probabilistic forecasts? What are desirable attributes of verification capabilities for probabilistic forecasts? What should a national verification capability encompass? What should a national verification capability encompass?

3 Verification terminology and links Verification is “The process of assessing the quality of forecasts” But – quality is closely linked to value Other words… EvaluationAssessment

4 Verification links Forecast verification is tightly linked (integral) to all aspects of forecast development, production, communication, and use Verification has significant impacts on Forecast development Forecast development Optimization of forecast models and systemsOptimization of forecast models and systems E.g., through choice of verification approaches, measures, and variables E.g., through choice of verification approaches, measures, and variables Post-processingPost-processing Presentation of forecast information Presentation of forecast information Decision making processes Decision making processes

5 Verification challenges Verification results should be understandable by the users of the information Verification results should be understandable by the users of the information Ex: What does a CSI of 0.28 really mean? Verification measures should appropriately distinguish and measure differences in performance Verification measures should appropriately distinguish and measure differences in performance Correlation = -0.02 POD = 0.00 FAR = 1.00 GSS (ETS) = -0.01 Correlation = 0.2 POD = 0.88 FAR = 0.89 GSS (ETS) = 0.08

6 Factors impacting evaluation of probabilistic forecasts Variable of interest Variable of interest Forecast type: Forecast type: DistributionDistribution ProbabilityProbability Point / spacePoint / space Verification attributes of interest Verification attributes of interest Reliability, Accuracy, DiscriminationReliability, Accuracy, Discrimination Spread-skillSpread-skill Timing errors; temporal changesTiming errors; temporal changes Spatial attributes (size, location, etc.)Spatial attributes (size, location, etc.) User of the verification information: User of the verification information: Forecast developer?Forecast developer? Forecaster?Forecaster? End user or decision making system?End user or decision making system? From NOAA/ESRL

7 Evaluation of distributions Ideal: Interpret ensemble in terms of a distribution Ideal: Interpret ensemble in terms of a distribution Many approaches for distribution fittingMany approaches for distribution fitting Many measures: Many measures: CRPS, CRPSSCRPS, CRPSS Rank Histogram (aka “Talagrand diagram”)Rank Histogram (aka “Talagrand diagram”) Minimum spanning tree (e.g., for cyclone forecasts?)Minimum spanning tree (e.g., for cyclone forecasts?) Probability measure (Wilson 1999)Probability measure (Wilson 1999) Ignorance scoreIgnorance score From Peel and Wilson 2008

8 Evaluation of probabilities Based on selection of meaningful threshold probabilities or events Based on selection of meaningful threshold probabilities or events Multiple measures provide information on accuracy, reliability, etc. Multiple measures provide information on accuracy, reliability, etc. Brier score, ROC, reliability, discriminationBrier score, ROC, reliability, discrimination Care is needed to appropriately select attributes of interest… and thresholds of interest Care is needed to appropriately select attributes of interest… and thresholds of interest

9 Treatment of Spatial Ensemble Forecasts As probabilities: Areas do not have “shape” of convective storms As mean: Area is not equivalent to any of the underlying ensemble members From C. Davis

10 Treatment of Spatial Ensemble Forecasts Alternative: Consider ensembles of “attributes”, such as… AreasAreas IntensitiesIntensities User-relevant measures (e.g., route blockages)User-relevant measures (e.g., route blockages) Collect and evaluate distributions of “attribute” errors

11 Links to value How do we measure value? How do we measure value? Richardson (2000): connect to ROCRichardson (2000): connect to ROC Case study approach: e.g., Keith (2005)Case study approach: e.g., Keith (2005) Econometric approaches (e.g., Lazo 2009)Econometric approaches (e.g., Lazo 2009) Connection between quality and value is not transparent Connection between quality and value is not transparent Richardson 2000Keith 2003

12 How do we meet the needs of diverse users? Provide information that is relevant to a wide spectrum of users Provide information that is relevant to a wide spectrum of users Ex: Multiple (user-selectable) thresholdsEx: Multiple (user-selectable) thresholds Evaluate a wide variety of forecast attributes Evaluate a wide variety of forecast attributes Focus on sensible weather elements Focus on sensible weather elements Utilize diagnostic techniques Utilize diagnostic techniques Ex: Distributions of statistics rather than (or in addition to) summary scoresEx: Distributions of statistics rather than (or in addition to) summary scores Provide access to “raw” forecasts and observations Provide access to “raw” forecasts and observations Ideally – strong interaction with users Ideally – strong interaction with users Understand spectrum of applications of verification informationUnderstand spectrum of applications of verification information Engage users in discussions of use and value of forecastsEngage users in discussions of use and value of forecasts Need to work with social scientists to understand information and communication needsNeed to work with social scientists to understand information and communication needs

13 Additional factors to consider… Sample size Sample size Dimensionality of probabilistic verification typically requires many cases for robust evaluationDimensionality of probabilistic verification typically requires many cases for robust evaluation Memory overload… Memory overload… Ensembles require a lot of storage and a lot of memory to process…Ensembles require a lot of storage and a lot of memory to process… Uncertainty in verification measures… Uncertainty in verification measures… MUST be shownMUST be shown How else do we make meaningful comparisons and measure progress in forecast development and improvement??? Or give honest appraisals of performance?How else do we make meaningful comparisons and measure progress in forecast development and improvement??? Or give honest appraisals of performance?

14 Summary Appropriate measurement and reporting of the quality of ensemble forecasts requires Understanding the integral nature of verification Understanding the integral nature of verification Measure and present forecast quality at all stagesMeasure and present forecast quality at all stages Consideration of the spectrum of forecast users and their interests and requirements Consideration of the spectrum of forecast users and their interests and requirements

15 Summary (cont.) Appropriate measurement and reporting of the quality of ensemble forecasts requires (cont…) Provision of diagnostic information that Provision of diagnostic information that Appropriately reflects the form and characteristics of the forecasts and observations (e.g., distribution; multiple probability thresholds)Appropriately reflects the form and characteristics of the forecasts and observations (e.g., distribution; multiple probability thresholds) Answers a wide variety of questions of interestAnswers a wide variety of questions of interest Measures a meaningful breadth of forecast performance attributes Measures a meaningful breadth of forecast performance attributes Easy access to this information for the entire community Easy access to this information for the entire community (Don’t keep this valuable information to ourselves) Contributions should be made by the entire community / enterprise

16

17 Recommendation 3.15: NWS should expand its verification systems for ensemble and other forecasts and make more explicit its choice of verification measures and rationale for those choices. Diagnostic and new verification approaches should be employed, and the verification should incorporate statistical standards such as stratification into homogeneous subgroups and estimation of uncertainty in verification measures. Verification information should be kept up to date and be easily accessible through the Web.

18 Recommendation 6. NWS should expand verification of its uncertainty products and make this information easily available to all users in near real time. A variety of verification measures and approaches (measuring multiple aspects of forecast quality that are relevant for users) should be used to appropriately represent the complexity and dimensionality of the verification problem. Verification statistics should be computed for meaningful subsets of the forecasts (e.g., by season, region) and should be presented in formats that are understandable by forecast users. Archival verification information on probabilistic forecasts, including model-generated and objectively generated forecasts and verifying observations, should be accessible so users can produce their own evaluation of the forecasts.


Download ppt "Verification and evaluation of a national probabilistic prediction system Barbara Brown NCAR 23 September 2009."

Similar presentations


Ads by Google