Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.

Slides:



Advertisements
Similar presentations
Chapter 3 Properties of Random Variables
Advertisements

Contingency Tables Chapters Seven, Sixteen, and Eighteen Chapter Seven –Definition of Contingency Tables –Basic Statistics –SPSS program (Crosstabulation)
OBJECTIVES Evaporation, precipitation and atmospheric heating ‘communicate’ SSTA to the atmosphere, driving changes in temperature, precipitation and.
Guidance of the WMO Commission for CIimatology on verification of operational seasonal forecasts Ernesto Rodríguez Camino AEMET (Thanks to S. Mason, C.
6th WMO tutorial Verification Martin GöberContinuous 1 Good afternoon! नमस्कार नमस्कार Guten Tag! Buenos dias! до́брый день! до́брыйдень Qwertzuiop asdfghjkl!
1 of Introduction to Forecasts and Verification.
Climate Predictability Tool (CPT)
Details for Today: DATE:3 rd February 2005 BY:Mark Cresswell FOLLOWED BY:Assignment 2 briefing Evaluation of Model Performance 69EG3137 – Impacts & Models.
Introduction to Probability and Probabilistic Forecasting L i n k i n g S c i e n c e t o S o c i e t y Simon Mason International Research Institute for.
Gridded OCF Probabilistic Forecasting For Australia For more information please contact © Commonwealth of Australia 2011 Shaun Cooper.
Statistical Tests Karen H. Hagglund, M.S.
Review of Basics. REVIEW OF BASICS PART I Measurement Descriptive Statistics Frequency Distributions.
Review of Basics. REVIEW OF BASICS PART I Measurement Descriptive Statistics Frequency Distributions.
Chapter Eighteen MEASURES OF ASSOCIATION
Chapter 13 Forecasting.
Barbara Casati June 2009 FMI Verification of continuous predictands
Chapter 7 Correlational Research Gay, Mills, and Airasian
Standard Scores & Correlation. Review A frequency curve either normal or otherwise is simply a line graph of all frequency of scores earned in a data.
Relationships Among Variables
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
Introduction to Statistics February 21, Statistics and Research Design Statistics: Theory and method of analyzing quantitative data from samples.
Creating Empirical Models Constructing a Simple Correlation and Regression-based Forecast Model Christopher Oludhe, Department of Meteorology, University.
Multi-Model Ensembling for Seasonal-to-Interannual Prediction: From Simple to Complex Lisa Goddard and Simon Mason International Research Institute for.
Introduction to Seasonal Climate Prediction Liqiang Sun International Research Institute for Climate and Society (IRI)
Seasonal Forecasting Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
Climate Predictability Tool (CPT) Ousmane Ndiaye and Simon J. Mason International Research Institute for Climate and Society The Earth.
Understanding Statistics
1 Measuring Association The contents in this chapter are from Chapter 19 of the textbook. The crimjust.sav data will be used. cjsrate: RATE JOB DONE: CJ.
Statistical analysis Prepared and gathered by Alireza Yousefy(Ph.D)
Applied Quantitative Analysis and Practices LECTURE#11 By Dr. Osman Sadiq Paracha.
Verification methods - towards a user oriented verification WG5.
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Heidke Skill Score (for deterministic categorical forecasts) Heidke score = Example: Suppose for OND 1997, rainfall forecasts are made for 15 stations.
Statistical analysis Outline that error bars are a graphical representation of the variability of data. The knowledge that any individual measurement.
Model dependence and an idea for post- processing multi-model ensembles Craig H. Bishop Naval Research Laboratory, Monterey, CA, USA Gab Abramowitz Climate.
Mathematics of PCR and CCA Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January.
Forecasting in CPT Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
Chapter 16 Data Analysis: Testing for Associations.
Climate Predictability Tool (CPT) Ousmane Ndiaye and Simon J. Mason International Research Institute for Climate and Society The Earth.
Can we distinguish wet years from dry years? Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand,
Evaluating Risk Adjustment Models Andy Bindman MD Department of Medicine, Epidemiology and Biostatistics.
Linear Regression Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015.
The Centre for Australian Weather and Climate Research A partnership between CSIRO and the Bureau of Meteorology Verification and Metrics (CAWCR)
Regression Analysis: Part 2 Inference Dummies / Interactions Multicollinearity / Heteroscedasticity Residual Analysis / Outliers.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Statistical analysis Why?? (besides making your life difficult …)  Scientists must collect data AND analyze it  Does your data support your hypothesis?
Chapter 16: Correlation. So far… We’ve focused on hypothesis testing Is the relationship we observe between x and y in our sample true generally (i.e.
1 Malaquias Peña and Huug van den Dool Consolidation of Multi Method Forecasts Application to monthly predictions of Pacific SST NCEP Climate Meeting,
©2005, Pearson Education/Prentice Hall CHAPTER 6 Nonexperimental Strategies.
Details for Today: DATE:13 th January 2005 BY:Mark Cresswell FOLLOWED BY:Practical Dynamical Forecasting 69EG3137 – Impacts & Models of Climate Change.
Verification methods - towards a user oriented verification The verification group.
Forecast 2 Linear trend Forecast error Seasonal demand.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Theme 5. Association 1. Introduction. 2. Bivariate tables and graphs.
Statistical analysis.
Non-Parametric Tests 12/1.
Statistical analysis.
Understanding the Relative Operating Characteristic (ROC)
Verifying and interpreting ensemble products
Question 1 Given that the globe is warming, why does the DJF outlook favor below-average temperatures in the southeastern U. S.? Climate variability on.
Climate Predictability Tool (CPT)
Relative Operating Characteristics
Basic Statistical Terms
Predictability of Indian monsoon rainfall variability
How good (or bad) are seasonal forecasts?
Seasonal Forecasting Using the Climate Predictability Tool
Can we distinguish wet years from dry years?
Forecast system development activities
Seasonal Forecasting Using the Climate Predictability Tool
Seasonal Forecasting Using the Climate Predictability Tool
Presentation transcript:

Model validation Simon Mason Seasonal Forecasting Using the Climate Predictability Tool Bangkok, Thailand, 12 – 16 January 2015

2 Seasonal Forecasting Using the Climate Predictability Tool Validation “Validation” v “verification”: we validate a model, but verify forecasts. In CPT, “validation” relates to the assessment of deterministic (“best guess”) cross-validated and retroactive predictions.

3 Seasonal Forecasting Using the Climate Predictability Tool Forecasts and observations DiscreteContinuous Deterministic It will rain tomorrow There will be 10 mm of rain tomorrow Probabilistic There is a 50% chance of rain tomorrow There is a p% chance of more than k mm of rain tomorrow

4 Seasonal Forecasting Using the Climate Predictability Tool The graph compares the cross-validated hindcasts (green) with the observed values (red). The observed climatological categories are shown in different colours. Are these good? Cross-validation of MAM rainfall for Thailand, using NIÑO4.

5 Seasonal Forecasting Using the Climate Predictability Tool Forecasts and observations DiscreteContinuous Deterministic It will rain tomorrow There will be 10 mm of rain tomorrow Probabilistic There is a 50% chance of rain tomorrow There is a p% chance of more than k mm of rain tomorrow

6 Seasonal Forecasting Using the Climate Predictability Tool Continuous measures compare the best-guess forecasts with the observed values without regard to the categories. They measure the skill of the deterministic forecasts. Tools ~ Validation ~ Cross-validated ~ Performance measures

7 Seasonal Forecasting Using the Climate Predictability Tool Pearson’s correlation Pearson’s correlation measures association (are increases and decreases in the forecasts associated with increases and decreases in the observations?). It does not measure accuracy. When squared, it tells us how much of the variance of the observations is correctly forecast.

8 Seasonal Forecasting Using the Climate Predictability Tool Pearson’s correlation For the Thailand forecasts, the correlation is 0.37, so about 14% of the variance of the observed rainfall was successfully predicted by the model. But measuring skill in terms of squared anomalies creates a few problems: 1. Sensitivity to extremes 2. Why are squared anomalies meaningful?

9 Seasonal Forecasting Using the Climate Predictability Tool Pearson’s correlation If we forecast March only, there are two extremely wet years (2001, with 280% of the average), which contributes to almost 25% of the variance in the observations. If one can forecast 2001 accurately, then 25% of the variance is already forecast.

10 Seasonal Forecasting Using the Climate Predictability Tool Spearman’s correlation Numerator:? Denominator:? How much of the squared variance of the ranks for the observations can we correctly forecast? Huh? Spearman’s correlation does not have as obvious an interpretation as Pearson’s, but it is much less sensitive to extremes.

11 Seasonal Forecasting Using the Climate Predictability Tool Kendall’s tau Denominator:total number of pairs. Numerator:difference in the numbers of concordant and discordant pairs. Kendall’s correlation measures discrimination (do the forecasts increase and decrease as the observations increase and decrease?). It can be transformed to the probability that the forecasts successfully distinguish the wetter (or hotter) of two observations?

12 Seasonal Forecasting Using the Climate Predictability Tool Kendall’s tau A concordant pair occurs when the largest of two X values corresponds with the largest of two Y values… In which of these two Januaries were the sea temperatures in the NINO3.4 region warmest (when did the stronger El Niño occur)? What is the probability of getting the answer correct? Repeat for all possible pairs of forecasts.

13 Seasonal Forecasting Using the Climate Predictability Tool Error measures compare the best-guess forecasts with the observed values without regard to the categories. They measure the skill of the deterministic forecasts.

14 Seasonal Forecasting Using the Climate Predictability Tool Biases indicate systematic errors in the forecasts.

15 Seasonal Forecasting Using the Climate Predictability Tool Biases Mean bias: Always close to zero for cross-validated forecasts; Indicates ability to forecast shifts in climate for retroactive forecasts; Slightly negative if predictand data are positively skewed. Variance or amplitude bias: Typically very small if skill is low because forecasts always close to the mean If there is no mean or variance bias, the RMSE of the forecasts will exceed that of climatology if the correlation is less than 0.5.

16 Seasonal Forecasting Using the Climate Predictability Tool Biases and GCMS ECHAM4.5 MAM rainfall for Thailand from Feb (purple).

17 Seasonal Forecasting Using the Climate Predictability Tool Biases and GCMS ECHAM4.5 rainfall for 21° – 27°N, 88° – 93°E (Bangladesh)

18 Seasonal Forecasting Using the Climate Predictability Tool Overall errors The MSE is a function of Pearson’s correlation, and the variance and mean biases. As with Pearson’s correlation, the MSE and RMSE are sensitive to the extremes because of the squaring. Mean absolute error is the most intuitive: on average the forecasts are “wrong” by about 55 mm.

19 Seasonal Forecasting Using the Climate Predictability Tool Forecasts and observations DiscreteContinuous Deterministic It will rain tomorrow There will be 10 mm of rain tomorrow Probabilistic There is a 50% chance of rain tomorrow There is a p% chance of more than k mm of rain tomorrow

20 Seasonal Forecasting Using the Climate Predictability Tool Categorical measures measure the skill of the deterministic forecasts with the observations as categories.

21 Seasonal Forecasting Using the Climate Predictability Tool Hit scores convert the forecasts to categories and then compare these with the observed categories. But remember that the category containing the best guess is not the most likely!

22 Seasonal Forecasting Using the Climate Predictability Tool Hit scores The contingency tables are based on cross-validated definitions of the categories and so may not perfectly match implied scores from the graph. Some hits can be expected even with useless forecasts (e.g., guessing, or always forecasting the same outcome… Tools ~ Contingency Tables ~ Cross-validated

23 Seasonal Forecasting Using the Climate Predictability Tool Hit skill scores Simply by guessing or always forecasting one category we would expect to get one-third of the forecasts correct (assuming the categories are equiprobable). The skill score counts how many more forecasts are “correct” than by guessing. Tools ~ Contingency Tables ~ Cross-validated

24 Seasonal Forecasting Using the Climate Predictability Tool Scoring near-misses Scoring one-category errors equally to two-category errors seems to be unsatisfactory. But penalizing for large errors has its own problems.

25 Seasonal Forecasting Using the Climate Predictability Tool Linear error in probability space (LEPS) These weights are defined to ensure that forecasts of climatology AND perpetual forecasts of one category AND random guessing have an expected score of zero. The score gives points for “near-misses”.

26 Seasonal Forecasting Using the Climate Predictability Tool Gerrity score This solution has some simpler properties.

27 Seasonal Forecasting Using the Climate Predictability Tool Measures of discrimination: can the forecasts successfully distinguish different outcomes? The observations are categories, but the forecasts are continuous (except where indicated).

28 Seasonal Forecasting Using the Climate Predictability Tool ROC areas: how frequently can the forecasts successfully distinguish different below-normal from normal and above- normal? Or above-normal from normal and below-normal?

29 Seasonal Forecasting Using the Climate Predictability Tool Discrimination – ROC areas When El Niño occurs, do we forecast warmer temperatures than when it does not occur?

30 Seasonal Forecasting Using the Climate Predictability Tool The same test can be extended to multiple categories – what is the probability that the observation in the higher category can be successfully identified.

31 Seasonal Forecasting Using the Climate Predictability Tool … and to infinite categories (or continuous values). What is the probability that we forecast a higher value when we observe a higher value?

32 Seasonal Forecasting Using the Climate Predictability Tool Continuous scores Correlations Pearson’s:% variance Spearman’s:% variance of ranks Kendall’s:2AFC – probability of successfully identifying warmer / wetter observation Errors Mean bias:unconditional error Variance bias:underestimation of variability RMSE:correlation, mean and variance bias MAE:average error

33 Seasonal Forecasting Using the Climate Predictability Tool Categorical scores Hits Hit score:% correct Hit skill:% correct adjusted for guessing LEPS:adjusts for near-misses Gerrity:adjusts for near-misses Discrimination 2AFC:probability of successfully identifying warmer / wetter category ROC:probability of successfully identifying observation in current category

34 Seasonal Forecasting Using the Climate Predictability Tool Significance testing Tools ~ Validation ~ Cross-validated ~ Bootstrap

35 Seasonal Forecasting Using the Climate Predictability Tool Exercises Assess the quality of your forecast model. Try to describe the quality of the model by interpreting the scores rather than just by listing them.