Presentation is loading. Please wait.

Presentation is loading. Please wait.

JAN | Feb-Mar-Apr Mar-Apr-May Apr-May-Jun May-Jun-Jul IRI’s monthly issued probability forecasts of seasonal global precipitation and temperature We issue.

Similar presentations


Presentation on theme: "JAN | Feb-Mar-Apr Mar-Apr-May Apr-May-Jun May-Jun-Jul IRI’s monthly issued probability forecasts of seasonal global precipitation and temperature We issue."— Presentation transcript:

1 JAN | Feb-Mar-Apr Mar-Apr-May Apr-May-Jun May-Jun-Jul IRI’s monthly issued probability forecasts of seasonal global precipitation and temperature We issue forecasts at four lead times. For example: Forecast models are run 7 months into future. Observed data are available through the end of the previous month (end of December in example above). Probabilities are given for the three tercile-based categories of the climatological distribution. “Now”| Forecasts made for:

2 0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 | || ||| ||||.| || | | || | | |. | | | | | | | | | | Rainfall Amount (mm) Below| Near | Below| Near | Above  Below| Near | The tercile category system: Below, near, and above normal (30 years of historical data for a particular location & season) Forecasts of the climate Data: 33% 33% 33% Probability:

3 0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Rainfall Amount (mm) Below| Near | Below| Near | Above  Below| Near | 20% 35% 45% Example of a typical climate forecast for a particular location & season Probability:

4 0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 Rainfall Amount (mm) Below| Near | Below| Near | Above  Below| Near | 5% 25% 70% Example of a STRONG climate forecast for a particular location & season Probability:

5 Below Above Historically, the probabilities of above and below are 0.33. Shifting the mean by half a standard-deviation and reducing the variance by 20% (red curve) changes the probability of below to 0.15 and of above to 0.53. Historical distribution Forecast distribution The probabilistic nature of climate forecasts

6 30 12 30 24 12 24 10 24 10 FORECAST SST TROP. PACIFIC: THREE (multi-models, dynamical and statistical) TROP. ATL, INDIAN (ONE statistical) EXTRATROPICAL (damped persistence) GLOBAL ATMOSPHERIC MODELS ECPC(Scripps) ECHAM4.5(MPI) CCM3.6(NCAR) NCEP(MRF9) NSIPP(NASA) COLA2 GFDL Forecast SST Ensembles 3/6 Mo. lead Persisted SST Ensembles 3 Mo. lead IRI DYNAMICAL CLIMATE FORECAST SYSTEM POST PROCESSING MULTIMODEL ENSEMBLING PERSISTED GLOBAL SST ANOMALY 2-tiered OCEAN ATMOSPHERE 30 GFDL has 10 to PSST

7 FORECAST SST TROP. PACIFIC: THREE scenarios: 1) CFS prediction 2) LDEO prediction 3) Constructed Analog prediction TROP. ATL, and INDIAN oceans CCA, or slowly damped persistence EXTRATROPICAL damped persistence MULTIPLE GLOBAL ATMOSPHERIC MODELS ECPC(Scripps) ECHAM4.5(MPI) CCM3.6(NCAR) NCEP(MRF9) NSIPP(NASA) COLA2 GFDL IRI DYNAMICAL CLIMATE FORECAST SYSTEM PERSISTED GLOBAL SST ANOMALY 2-tiered OCEAN ATMOSPHERE

8 Atmospheric General Circulation Models Used in the IRI's Seasonal Forecasts, for Superensembles Name Where Model Was Developed Where Model Is Run NCEP MRF-9 NCEP, Washington, DC QDNR, Queensland, Australia ECHAM 4.5 MPI, Hamburg, Germany IRI, Palisades, New York NSIPP NASA/GSFC, Greenbelt, MD NASA/GSFC, Greenbelt, MD COLA COLA, Calverton, MD COLA, Calverton, MD ECPC SIO, La Jolla, CA SIO, La Jolla, CA CCM3.6 NCAR, Boulder, CO IRI, Palisades, New York GFDL GFDL, Princeton, NJ GFDL, Princeton, NJ Collaboration on Input to Forecast Production Sources of the Global Sea Surface Temperature Forecasts Tropical PacificTropical AtlanticIndian OceanExtratropical Oceans NCEP CoupledCPTEC Statistical IRI StatisticalDamped Persistence LDEO Coupled Constr Analogue

9 ----- Much of our predictability comes about due to

10

11 Strong trade winds Westward currents, upwelling Cold east, warm west Convection, rising motion in west Weak trade winds Eastward currents, suppressed upwelling Warm west and east Enhanced convection, eastward displacement

12 NDJ Precipitation Probabilities associated with El Nino Empirical Approach (a good baseline by which to judge the more advanced methods “honest”) NDJ

13 Combining the Predictions of Seasonal Climate by Several Atmospheric General Circulation Models (AGCMs) into a Single Prediction. What is Done at IRI

14 Goals in Combining To combine the probability forecasts of several models, with relative weights based on the past performance of the individual models To assign appropriate forecast probability distribution: e.g. damp overconfident forecasts toward climatology

15 One Choice: Multiple Regression Multiple regression uses the predictions of 2 or more models, along with the corresponding observations, to derive a set of model weights that minimizes squared errors of the weighted predictions. Good feature: The actual skills of each model are taken into account, and also the overlap in sources of predictability. So, if two models have high skill in the same way or for the same reason, or if two good models are nearly identical, both of them do not get the same high weight. Bad features: (1) Random variations in the training sample are taken as seriously as the predictable part of the variability. This leads to overfitting, and poor results in real forecasts. The more predictors compared to the number of independent cases, the more severe this problem is. (2) Forecasts for conditions outside the range of the training sample may be dangerous. Note: A model’s weight does not necessarily describe its skill. In fact, negative weights can be seen for skillful models.

16 Another Choice: Simple Skill Weighting Here, the weights given to 2 or more models are determined by each model’s individual skill in forecasting observations. The skill may be a correlation or other measure. The weights are normalized so that their sum is 1. Good features: (1) The actual skills of each model are taken into account. (2) Overfitting is not as serious a problem here as it is in multiple regression. (3) Final forecast is always within the range of the individual forecasts. Bad feature: Duplication in the sources of skill is not checked. (Colinearity ignored.) So, if two good models are almost identical, both will be weighted strongly, even though the second one does not add much skill. A medium-skill model that has a unique source of skill will not influence the forecast enough. Note: A model’s weight is easily interpreted as being its skill.

17 Combine climatology forecast (“prior”) and an AGCM forecast, with its evidence of historical skill, to produce weighted (“posterior”) forecast probabilities, by maximizing the historical likelihood score. Methods Used at IRI: (1) Bayesian Combination (2) Canonical Variate Combination Bayesian Combination

18 Six GCM Precip. Forecasts, JAS 2000

19 RPSS Skill of Individual Models: JAS 1950-97

20 Probabilities and Uncertainty 1/3 6/24 8/24 10/24 Above Normal Near-Normal Below Normal Climatological Probabilities GCM Probabilities Tercile boundaries are identified for the models’ own climatology, by aggregating all years and ensemble members. This corrects overall bias. k = tercile number t = forecast time m = no. ens members

21 Bayesian Model Combination Combine climatology forecast (“prior”) and an AGCM forecast, with its evidence of historical skill, to produce weighted (“posterior”) forecast probabilities, by maximizing the historical likelihood score.

22 The multi-year product of the probabilities that were hindcast for the category that was observed. (Could maximize other scores, such as RPSS) Prescribed, observed SST used to force AGCMs. Such simulations used in absence of ones using truly forecasted SST for at least half of AGCMs. Aim to maximize the likelihood score k=tercile category t=year number

23 1. Calibration of each model, individually, against climatology 2. Calibration of the weighted model combination against climatol k=tercile category (1,2, or 3) t=year number j=model number (1 to 7) w=weight for climo (c) or for model j where w MM uses w j proportional to results of the first step above P MMkt = weighted linear comb of P jkt over all j, normalized by Σ(w j ) Optimize likelihood score Optimize likelihood score

24 Algorithm used to maximize the designated score: Feasible Sequential Quadratic Programming (FSQP) “Nonmonotone line search for minimax problems” C M : TOTAL NUMBER OF CONSTRAINTS. C ME : NUMBER OF EQUALITY CONSTRAINTS. C MMAX : ROW DIMENSION OF A. MMAX MUST BE AT LEAST ONE AND GREATER C THAN M. C N : NUMBER OF VARIABLES. C NMAX : ROW DIMENSION OF C. NMAX MUST BE GREATER OR EQUAL TO N. C MNN : MUST BE EQUAL TO M + N + N. C C(NMAX,NMAX): OBJECTIVE FUNCTION MATRIX WHICH SHOULD BE SYMMETRIC AND C POSITIVE DEFINITE. IF IWAR(1) = 0, C IS SUPPOSED TO BE THE C CHOLESKEY-FACTOR OF ANOTHER MATRIX, I.E. C IS UPPER C TRIANGULAR. C D(NMAX) : CONTAINS THE CONSTANT VECTOR OF THE OBJECTIVE FUNCTION. C A(MMAX,NMAX): CONTAINS THE DATA MATRIX OF THE LINEAR CONSTRAINTS. C B(MMAX) : CONTAINS THE CONSTANT DATA OF THE LINEAR CONSTRAINTS. C XL(N),XU(N): CONTAIN THE LOWER AND UPPER BOUNDS FOR THE VARIABLES. C X(N) : ON RETURN, X CONTAINS THE OPTIMAL SOLUTION VECTOR. C U(MNN) : ON RETURN, U CONTAINS THE LAGRANGE MULTIPLIERS. THE FIRST C M POSITIONS ARE RESERVED FOR THE MULTIPLIERS OF THE M C LINEAR CONSTRAINTS AND THE SUBSEQUENT ONES FOR THE C MULTIPLIERS OF THE LOWER AND UPPER BOUNDS. ON SUCCESSFUL C TERMINATION, ALL VALUES OF U WITH RESPECT TO INEQUALITIES C AND BOUNDS SHOULD BE GREATER OR EQUAL TO ZERO.

25 Circumventing the effects of sampling variability Sampling variability appears to be an issue: noisy weight distribution with large number of zero weights and some unity weights Bootstrap the optimization, omitting contiguous 6-year blocks of the 48-yr time series –yields 43 samples of 42 years –shows the sampling variability of the likelihood over subsets of years –We average the weights across the samples

26 Example Six GCMs’ Jul-Aug-Sep precipitation simulations Training period: 1950–97 Ensembles of between 9 and 24 members

27 Model Weights – initially, by individual model

28 Climatological Weights – Multi- model

29 Model Weights – after second (damping) step

30 Model Weights – step 2, and Averaged over Subsamples

31 For more spatially smooth results, the weighting of each grid point is averaged with that of its 8 neighbors, using binomial weighting. X X X

32 Climatological Weights

33 Combination Forecasts of July-Sept Precipitation After first stage onlyAfter second (damping) stage After spatial smoothingAfter sampling subperiods

34 Reliability JAS Precip., 30S-30N Above-NormalBelow-Normal Forecast probability Observed relative Freq. Bayesian Pooled (3-model) from Goddard et al. 2003 Individual AGCM

35 RPSS Precip. from Roberson et al. (2004): Mon. Wea. Rev., 132, 2732-2744

36 RPSS 2-m Temp. from Roberson et al. (2004): Mon. Wea. Rev., 132, 2732-2744

37 Conclusions - Bayesian The “climatological” (equal-odds) forecast provides a useful prior for combining multiple ensemble forecasts Sampling problems become severe when attempting to combine many models from a short training period (“noisy weights”) A two-stage process combines the models together according to a pre-assessment of each against climatology Smoothing of the weights across data sub- samples and spatially appears beneficial

38 IRI’s forecasts use also a second consolidation scheme, whose result is averaged with the result of the Bayesian scheme. 1. Bayesian scheme 2. Canonical Variate scheme

39 Canonical Variate Analysis (CVA) A number of statistical techniques involve calculating linear combinations (weighted sums) of variables. The weights are defined to achieve specific objectives: PCA – weighted sums maximize variance CCA – weighted sums maximize correlation CVA – weighted sums maximize discrimination

40 Canonical Variate Analysis

41 The canonical variates are defined to maximize the ratio of the between-category (separation between the crosses) to the within-category (separation of dots from like-colored crosses) variance.

42 Conclusion IRI is presently using a 2-tiered prediction system. It is interested in using fully coupled systems also, and is looking into incorporating those. Within its 2-tiered system it uses 4 SST prediction scenarios, and combines the predictions of 7 AGCMs. The merging of 7 predictions into a single one uses two multi-model ensemble systems: Bayesian and canonical variate. These give somewhat differing solutions, and are presently given equal weight.

43 Plan for Improvement of the Basic Forecast Product IRI plans to issue forecasts as probability density functions (pdfs) instead of just for the 3 tercile-based categories. From the pdfs, users can construct proba- bilities for any categories desired.

44 Skill results for IRI real-time climate forecasts from 1997-2001

45 Real-time Forecast Skill Goddard et al. 2003, Bull. Amer. Meteorol. Soc., 84, 1973-1796.

46 Real-time Forecast Skill Goddard et al. 2003, Bull. Amer. Meteorol. Soc., 84, 1973-1796.

47 Real-time Forecast Skill Goddard et al. 2003, Bull. Amer. Meteorol. Soc., 84, 1973-1796.

48 Real-time Forecast Skill Goddard et al. 2003, Bull. Amer. Meteorol. Soc., 84, 1973-1796.

49

50 Perfect reliability RELIABILITY DIAGRAM

51 Climate Information System: A USER centric perspective Issues about relating to our users

52 Some Common Customer Requests 1.More accurate climate forecasts (tighter pdfs) (we can at least deliver most reliable ones) 2. Forecasts of full pdf in physical units 3. Forecasts of the weather within climate 4. Forecasts for subseasonal periods 5. Climate change nowcast


Download ppt "JAN | Feb-Mar-Apr Mar-Apr-May Apr-May-Jun May-Jun-Jul IRI’s monthly issued probability forecasts of seasonal global precipitation and temperature We issue."

Similar presentations


Ads by Google