Presentation is loading. Please wait.

Presentation is loading. Please wait.

Atmospheric Sciences 452 Spring 2019

Similar presentations


Presentation on theme: "Atmospheric Sciences 452 Spring 2019"— Presentation transcript:

1 Atmospheric Sciences 452 Spring 2019
Model Post Processing Atmospheric Sciences 452 Spring 2019

2 Model Output Can Usually Be Improved with Statistical Post Processing
Can remove systematic bias Can produce probabilistic information from deterministic information and historical performance. Can provide forecasts for parameters that a model is incapable of simulating successfully due to resolution or physics issues (e.g., shallow fog)

3 Model Post-Processing
There are a variety of approaches: Simple bias removal Model output statistics (MOS) Machine learning and artificial intelligence (AI) Bayesian model averaging Neutral nets And more….

4 Model Output Statistics (MOS)
Model Output Statistics was the first post-processing method used by the NWS (1969) Based on multiple linear regression. Essentially unchanged over past 40 years. Does not consider non-linear relationships between predictors and predictands. Does take out much of systematic bias. Does improve forecast!

5 Based on Multiple Linear Regression
Y=a0 + a1 X1+ a2X2 + … Y is the predictand—what we want to predict Xi are the predictors….can be model output or observations ai are coefficients

6 Multiple Linear Regression in MOS
Select Xi and calculate ai using BOTH model and observational data. So Xi and be either model output or observations. Advantage: can adjust for model biases (e.g., too warm) Disadvantage: Needs several years of model runs to derive equations. If model changes significantly, have to do it again.

7 How do we select Xi? Use screening regression
Start with at least two years of model output and observations. Go through a long laundry list of potential predictors Select one that has the highest correlation with the predictand. Then select a second predictor that has produces the most increase in correlation in addition to the first predictor. Keep on going, typically until 12 predictors are found.

8 Correlation Coefficient (r) Between Two Time series
0—no correlation 1—perfect correlation R2 = percent of explained variance So r = .5, 25% of the variance or variability of the second time series can be explained by the variability in the first.

9

10 Why does the NWS stop at 12 predictors?
Too many predictors can be detrimental! Adds little improvement but adds to computational expense Can overfit the data/ Why? Because one can fit small, random, or unique events. Most of the benefit are with the first 3-4 predictors. Generally, takes at least a two year sample of forecasts to have a sufficient data base.

11 Examples!

12 Day 2 (30-h) GFS MOS Max Temp Equation for KSLC (Cool Season – 0000 UTC cycle)
Predictor (XN) Coeff. (aN) 1 2-m Temperature (21-h proj.) 1.7873 2 2-m Dewpoint (21-h proj.) 0.1442 3 2-m Dewpoint (12-h proj.) 4 2-m Dewpoint (27-h proj.) 0.1252 5 Observed Temperature (03Z) 0.0354 6 850-mb Vertical Velocity (21-h proj.) 7 925-mb Wind Speed (15-h proj.) 0.6024 8 Sine Day of Year 1.7111 9 700-mb Wind Speed (15-h proj.) 0.2701 10 Sine 2*DOY 1.5110

13 Day 2 (42-h) GFS MOS Max Temp Equation for KUNV (Warm Season – 1200 UTC cycle)
Predictor (XN) Coeff. (aN) 1 2-m Temperature (33-h proj.) 0.9249 2 2-m Dewpoint (33-h proj.) 0.5751 3 950-mb Dewpoint (24-h proj.) 0.4026 4 950-mb Rel. Humidity (27-h proj.) 5 850-mb Dewpoint (39-h proj.) 6 Observed Dewpoint (15Z) 7 Observed Temperature (15Z) 0.1270 8 1000-mb Rel. Humidity (24-h proj.) 0.0027 9 Sine Day of Year 0.9763 10 mb Thickness (45-h proj.) 0.0057

14 Day 2 (30-h) GFS MOS Min Temp Equation for KDCA (Cool Season - 1200 UTC cycle)
Predictor (XN) Coeff. (aN) 1 2-m Temperature (21-h proj.) 0.9700 2 1000-mb Temperature (12-h proj.) 0.3245 3 2-m Dewpoint (27-h proj.) 0.1858 4 2-m Relative Humidity (27-h proj.) 5 2-m Relative Humidity (15-h proj.) 6 975-mb Wind Speed (21-h proj.) 7 Observed Temperature (15Z) 0.1584 8 Sine Day of Year Develop / Evaluate

15

16 There are HUNDREDS of THOUSANDS of MOS equations
Separate equations for each of thousands of locations (see more on this later) Different equations for each variable Different equations for each season, initialization time (e.g., 0 or 12 UTC) Different equation for each projection (e.g., 30 h forecasts)

17 A Very Important Point For some (most) parameters, a different equation for each station. Parameters in which there samples each hour (e.g., temperature, dew point) Such equations can take in consideration local biases, local weather features, etc. Thus, temperature MOS knows about the Puget Sound convergence zone even if the model doesn’t.

18 A Very Important Point For other (most) parameters, the same equation for nearby station. Parameters in which there FEW samples each hour (e.g., precipitation, severe weather) Such equations DO NOT take in consideration local biases, local weather features, etc. Thus, these MOS equations do no add any information on local weather features

19

20

21

22 MOS Characteristics Tends to go toward climatology at longer periods
Does poorly for transient (short-term) model failures From poor synoptic forecast Unusual biases Can go for extreme events, but often misses them Only some MOS parameters can add local impacts.

23 MOS Comments Good for “garden variety” events. Very hard for humans to beat—but it is possible for some situations (shallow cold air) Average of several MOS forecasts (e.g., NAM and GFS), better than single MOS MOS reduces or removes long-term, systematic biases. Does little for rare or transient biases.

24 More MOS Available for several modeling systems: NAM MOS GFS MOS
GEFS MOS

25 MOS Developed by and Run at the NWS Meteorological Development Lab (MDL)
Full range of products available at:

26

27

28 Global Ensemble MOS Ensemble MOS forecasts are based on the 0000 UTC run of the GFS Global model ensemble system. These runs include the operational GFS, a control version of the GFS (run at lower resolution), and 20 additional runs. Older operational GFS MOS prediction equations are applied to the output from each of the ensemble runs to produce 21 separate sets of alphanumeric bulletins in the same format as the operational MEX message.

29 Gridded MOS The NWS needs MOS on a grid for many reasons, including for use in their IFPS analysis/forecasting system. The problem is that MOS is only available at station locations. To deal with this, NWS created Gridded MOS. Takes MOS at individual stations and spreads it out based on proximity and height differences. Also does a topogaphic correction dependent on reasonable lapse rate.

30

31

32 Current “Operational” Gridded MOS

33 Localized Aviation MOS Program (LAMP)
Hourly updated statistical product Like MOS but combines: MOS guidance the most recent surface observations simple local models run hourly GFS output

34 Practical Example of Solving a LAMP Temperature Equation
Y = b + a1x1 + a2x2 + a3x3 + a4x4 Y = LAMP temperature forecast Equation Constant b = Predictor x1 = observed temperature at cycle issuance time (value 66.0) Predictor x2 = observed dew point at cycle issuance time (value 58.0) Predictor x3 = GFS MOS temperature (value 64.4) Predictor x4 = GFS MOS dew point (value 53.0)

35 Theoretical Model Forecast Performance of LAMP, MOS, and Persistence

36 Verification of LAMP 2-m Temperature Forecasts

37 MOS Performance Versus Humans
MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between human and MOS forecasts.

38 Cool Season Mi. Temp – 12 UTC Cycle
Average Over 80 US stations

39 MOS Won the Department Forecast Contest in 2003 For the First Time!

40

41

42

43 UW MOS Study August 1 2003 – August 1 2004 (12 months).
29 stations, all at major NWS Weather Forecast Office (WFO) sites. Evaluated MOS predictions of maximum and minimum temperature, and probability of precipitation (POP).

44 National Weather Service locations used in the study.

45 Forecasts Evaluated NWS Forecast by real, live humans EMOS: NAM MOS
NMOS: NGM MOS GMOS: GFS MOS CMOS: Average of the above three MOSs WMOS: Weighted MOS, each member is weighted by its performance during a previous training period (ranging from days, depending on each station). CMOS-GE: A simple average of the two best MOS forecasts: GMOS and EMOS

46 The Approach: Give the NWS the Advantage!
08-10Z-issued forecast from NWS matched against previous 00Z forecast from models/MOS. NWS has 00Z model data available, and has added advantage of watching conditions develop since 00Z. Models of course can’t look at NWS, but NWS looks at models. NWS Forecasts going out 48 (model out 60) hours, so in the analysis there are: Two maximum temperatures (MAX-T), Two minimum temperatures (MIN-T), and Four 12-hr POP forecasts.

47 Temperature Comparisons

48 Temperature MAE (F) for the seven forecast types for all stations, all time periods, 1 August – 1 August 2004.

49 Precipitation Comparisons

50 Brier Scores for Precipitation for all stations for the entire study period.

51 New NWS Approach: National Blend of Models
The National Blend of Models (NBM) is a nationally consistent and skillful suite of calibrated forecast guidance based on a blend of both NWS and non-NWS numerical weather prediction model data and post-processed model guidance. The goal of the NBM is to create a highly accurate, skillful and consistent starting point for the gridded forecast.

52 Blend Combines statistically a collection of models 2.5 km grid
website

53

54

55

56 National Blend of Models
Will be the starting point of NWS gridded forecasts Save forecasters time Reduce issues between boundaries of different offices.

57 The Private Sector Has Gone Beyond MOS to Superior Post Processing

58 They don’t do traditional MOS!

59 Dynamic MOS Using Multiple Models
MOS equations are updated frequently, not static like the NWS. Multiple model and observation inputs Example: DiCast used by the Weather Channel and Accuweather, Developed by NCAR

60 Dynamical MOS of MOSs

61 DICAST skill is quite good

62 ForecastAdvisor.com

63

64 Better than NWS MOS

65 Bayesian Model Averaging (BMA)
A good way to optimize the use of ensembles of forecasts to provide calibrated probabilistic guidance. Can weight models and the variability by previous performance.

66

67

68 Machine Learning/Artificial Intelligence

69 Machine learning algorithms use computational methods to “learn” information directly from data without relying on a predetermined equation The algorithms adaptively improve their performance as the number of samples available for learning increases. Machine learning is a field of computer science that uses statistical techniques to give computer systems the ability to "learn" (e.g., progressively improve performance on a specific task) with data, without being explicitly programmed.

70 Two Types of Machine Learning

71

72 Decision Trees are Very Popular Particularly Random Forest

73 Neural nets Attempts to duplicate the complex interactions between neurons in the human brain.

74


Download ppt "Atmospheric Sciences 452 Spring 2019"

Similar presentations


Ads by Google