Model Post Processing.

Slides:



Advertisements
Similar presentations
Chapter 13 – Weather Analysis and Forecasting
Advertisements

A Brief Guide to MDL's SREF Winter Guidance (SWinG) Version 1.0 January 2013.
KMA will extend medium Range forecast from 7day to 10 day on Oct A post processing technique, Ensemble Model Output Statistics (EMOS), was developed.
Part 5. Human Activities Chapter 13 Weather Forecasting and Analysis.
Gridded OCF Probabilistic Forecasting For Australia For more information please contact © Commonwealth of Australia 2011 Shaun Cooper.
1 Operational low visibility statistical prediction Frédéric Atger (Météo-France)
Similar Day Ensemble Post-Processing as Applied to Wildfire Threat and Ozone Days Michael Erickson 1, Brian A. Colle 1 and Joseph J. Charney 2 1 School.
PERFORMANCE OF NATIONAL WEATHER SERVICE FORECASTS VERSUS MODEL OUTPUT STATISTICS Jeff Baars Cliff Mass Mark Albright University of Washington, Seattle,
Reliability Trends of the Global Forecast System Model Output Statistical Guidance in the Northeastern U.S. A Statistical Analysis with Operational Forecasting.
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Hydrometeorological Prediction Center HPC Medium Range Grid Improvements Mike Schichtel, Chris Bailey, Keith Brill, and David Novak.
MOS What does acronym stand for ? –MODEL OUTPUT STATISTICS What is the difference between the GFS and GFS MOS ?
JEFS Status Report Department of Atmospheric Sciences University of Washington Cliff Mass, Jeff Baars, David Carey JEFS Workshop, August
MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between.
Ensemble Post-Processing and it’s Potential Benefits for the Operational Forecaster Michael Erickson and Brian A. Colle School of Marine and Atmospheric.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
Weather and climate monitoring for food risk management G. Maracchi WMO, Geneva, November 2004 Weather and climate monitoring for food risk management.
Lecture 10 (11/11) Numerical Models. Numerical Weather Prediction Numerical Weather Prediction (NWP) uses the power of computers and equations to make.
SRNWP workshop - Bologne Short range ensemble forecasting at Météo-France status and plans J. Nicolau, Météo-France.
Forecasting and Numerical Weather Prediction (NWP) NOWcasting Description of atmospheric models Specific Models Types of variables and how to determine.
Toward a 4D Cube of the Atmosphere via Data Assimilation Kelvin Droegemeier University of Oklahoma 13 August 2009.
Exploring sample size issues for 6-10 day forecasts using ECMWF’s reforecast data set Model: 2005 version of ECMWF model; T255 resolution. Initial Conditions:
June 19, 2007 GRIDDED MOS STARTS WITH POINT (STATION) MOS STARTS WITH POINT (STATION) MOS –Essentially the same MOS that is in text bulletins –Number and.
By John Metz Warning Coordination Meteorologist WFO Corpus Christi.
Celeste Saulo and Juan Ruiz CIMA (CONICET/UBA) – DCAO (FCEN –UBA)
OUTLINE Current state of Ensemble MOS
MDL Lightning-Based Products Kathryn Hughes NOAA/NWS/OST December 3, 2003
Model Post Processing. Model Output Can Usually Be Improved with Post Processing Can remove systematic bias Can produce probabilistic information from.
CC Hennon ATMS 350 UNC Asheville Model Output Statistics Transforming model output into useful forecast parameters.
Application of an adaptive radiative transfer parameterisation in a mesoscale numerical weather prediction model DWD Extramural research Annika Schomburg.
Production of a multi-model, convective- scale superensemble over western Europe as part of the SESAR project EMS Annual Conference, Sept. 13 th, 2013.
The LAMP/HRRR MELD FOR AVIATION FORECASTING Bob Glahn, Judy Ghirardelli, Jung-Sun Im, Adam Schnapp, Gordana Rancic, and Chenjie Huang Meteorological Development.
MOS AVN = Dynamical Model –Seven fundamental equations ! AVN MOS = Statistical Model –No seven fundamental equations ! –Equations are statistical, not.
Kris Shrestha James Belanger Judith Curry Jake Mittelman Phillippe Beaucage Jeff Freedman John Zack Medium Range Wind Power Forecasts for Texas.
Overview of WG5 activities and Conditional Verification Project Adriano Raspanti - WG5 Bucharest, September 2006.
MDL Requirements for RUA Judy Ghirardelli, David Myrick, and Bruce Veenhuis Contributions from: David Ruth and Matt Peroutka 1.
Judith Curry James Belanger Mark Jelinek Violeta Toma Peter Webster 1
Translating Advances in Numerical Weather Prediction into Official NWS Forecasts David P. Ruth Meteorological Development Laboratory Symposium on the 50.
Developing GFS-based MOS Thunderstorm Guidance for Alaska Phillip E. Shafer* and Kathryn Gilbert Meteorological Development Laboratory, NWS, NOAA
National Oceanic and Atmospheric Administration’s National Weather Service Colorado Basin River Forecast Center Salt Lake City, Utah 11 The Hydrologic.
Chapter 13 – Weather Analysis and Forecasting. The National Weather Service The National Weather Service (NWS) is responsible for forecasts several times.
LAMP improvements for ceiling height and visibility guidance*
Encast Global forecasting.
Medium Range Forecasting at the Weather Prediction Center (WPC) –
A Guide to Tropical Cyclone Guidance
Overview of Downscaling
Realities (NDFD, NDGD, IFPS)
Update on the Northwest Regional Modeling System 2013
University of Washington Ensemble Systems for Probabilistic Analysis and Forecasting Cliff Mass, Atmospheric Sciences University of Washington.
Overview of Deterministic Computer Models
Tom Hopson, Jason Knievel, Yubao Liu, Gregory Roux, Wanli Wu
Bo Christiansen Downward propagation from the stratosphere:
Winter storm forecast at 1-12 h range
Update on the Northwest Regional Modeling System 2017
MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at:
Improving forecasts through rapid updating of temperature trajectories and statistical post-processing Nina Schuhen, Thordis L. Thorarinsdottir and Alex.
Post Processing.
Post Processing.
452 Precipitation.
The Importance of Reforecasts at CPC
WMO NWP Wokshop: Blending Breakout
Andy Wood and Dennis P. Lettenmaier
New Developments in Aviation Forecast Guidance from the RUC
Ensemble-4DWX update: focus on calibration and verification
Alex Gallagher and Dr. Robert Fovell
MOS What does acronym stand for ?
Atmospheric Sciences 452 Spring 2019
Model Output Statistics
Rapid Adjustment of Forecast Trajectories: Improving short-term forecast skill through statistical post-processing Nina Schuhen, Thordis L. Thorarinsdottir.
Short Range Ensemble Prediction System Verification over Greece
Presentation transcript:

Model Post Processing

Model Output Can Usually Be Improved with Post Processing Can remove systematic bias Can produce probabilistic information from deterministic information and historical performance. Can provide forecasts for parameters that a model is incapable of simulating successfully due to resolution or physics issues (e.g., shallow fog)

Model Output Statistics (MOS) Model Output Statistics was the first post-processing method used by the NWS (1969) Based on multiple linear regression. Essentially unchanged over 40 years. Does not consider non-linear relationships between predictors and predictands. Does take out much of systematic bias.

Based on Multiple Linear Regression Y=a0 + a1 X1+ a2X2 + …

Day 2 (30-h) GFS MOS Max Temp Equation for KSLC (Cool Season – 0000 UTC cycle) Predictor (XN) Coeff. (aN) -467.7800 1 2-m Temperature (21-h proj.) 1.7873 2 2-m Dewpoint (21-h proj.) 0.1442 3 2-m Dewpoint (12-h proj.) -0.2060 4 2-m Dewpoint (27-h proj.) 0.1252 5 Observed Temperature (03Z) 0.0354 6 850-mb Vertical Velocity (21-h proj.) 19.5759 7 925-mb Wind Speed (15-h proj.) 0.6024 8 Sine Day of Year 1.7111 9 700-mb Wind Speed (15-h proj.) 0.2701 10 Sine 2*DOY 1.5110

Day 2 (42-h) GFS MOS Max Temp Equation for KUNV (Warm Season – 1200 UTC cycle) Predictor (XN) Coeff. (aN) -432.2469 1 2-m Temperature (33-h proj.) 0.9249 2 2-m Dewpoint (33-h proj.) 0.5751 3 950-mb Dewpoint (24-h proj.) 0.4026 4 950-mb Rel. Humidity (27-h proj.) -0.1784 5 850-mb Dewpoint (39-h proj.) -0.2439 6 Observed Dewpoint (15Z) -0.1000 7 Observed Temperature (15Z) 0.1270 8 1000-mb Rel. Humidity (24-h proj.) 0.0027 9 Sine Day of Year 0.9763 10 500-1000mb Thickness (45-h proj.) 0.0057

Day 2 (30-h) GFS MOS Min Temp Equation for KDCA (Cool Season - 1200 UTC cycle) Predictor (XN) Coeff. (aN) -374.9980 1 2-m Temperature (21-h proj.) 0.9700 2 1000-mb Temperature (12-h proj.) 0.3245 3 2-m Dewpoint (27-h proj.) 0.1858 4 2-m Relative Humidity (27-h proj.) -0.0037 5 2-m Relative Humidity (15-h proj.) -0.0380 6 975-mb Wind Speed (21-h proj.) -0.0653 7 Observed Temperature (15Z) 0.1584 8 Sine Day of Year -0.0342 Develop / Evaluate

MOS Developed by and Run at the NWS Meteorological Development Lab (MDL) Full range of products available at: http://www.nws.noaa.gov/mdl/synop/index.php

Global Ensemble MOS Ensemble MOS forecasts are based on the 0000 UTC run of the GFS Global model ensemble system. These runs include the operational GFS, a control version of the GFS (run at lower resolution), and 20 additional runs. Older operational GFS MOS prediction equations are applied to the output from each of the ensemble runs to produce 21 separate sets of alphanumeric bulletins in the same format as the operational MEX message.

Gridded MOS The NWS needs MOS on a grid for many reasons, including for use in their IFPS analysis/forecasting system. The problem is that MOS is only available at station locations. To deal with this, NWS created Gridded MOS. Takes MOS at individual stations and spreads it out based on proximity and height differences. Also does a topogaphic correction dependent on reasonable lapse rate.

Current “Operational” Gridded MOS

Localized Aviation MOS Program (LAMP) Hourly updated statistical product Like MOS but combines: MOS guidance the most recent surface observations simple local models run hourly GFS output

Practical Example of Solving a LAMP Temperature Equation Y = b + a1x1 + a2x2 + a3x3 + a4x4 Y = LAMP temperature forecast Equation Constant b = -6.99456 Predictor x1 = observed temperature at cycle issuance time (value 66.0) Predictor x2 = observed dew point at cycle issuance time (value 58.0) Predictor x3 = GFS MOS temperature (value 64.4) Predictor x4 = GFS MOS dew point (value 53.0)

Theoretical Model Forecast Performance of LAMP, MOS, and Persistence

Verification of LAMP 2-m Temperature Forecasts

MOS Performance MOS significantly improves on the skill of model output. National Weather Service verification statistics have shown a narrowing gap between human and MOS forecasts.

Cool Season Mi. Temp – 12 UTC Cycle Average Over 80 US stations

MOS Won the Department Forecast Contest in 2003 For the First Time!

Average or Composite MOS There has been some evidence that an average or consensus MOS is even more skillful than individual MOS output. Vislocky and Fritsch (1997), using 1990-1992 data, found that an average of two or more MOS’s (CMOS) outperformed individual MOS’s and many human forecasters in a forecasting competition.

UW MOS Study August 1 2003 – August 1 2004 (12 months). 29 stations, all at major NWS Weather Forecast Office (WFO) sites. Evaluated MOS predictions of maximum and minimum temperature, and probability of precipitation (POP).

National Weather Service locations used in the study.

Forecasts Evaluated NWS Forecast by real, live humans EMOS: Eta MOS NMOS: NGM MOS GMOS: GFS MOS CMOS: Average of the above three MOSs WMOS: Weighted MOS, each member is weighted by its performance during a previous training period (ranging from 10-30 days, depending on each station). CMOS-GE: A simple average of the two best MOS forecasts: GMOS and EMOS

The Approach: Give the NWS the Advantage! 08-10Z-issued forecast from NWS matched against previous 00Z forecast from models/MOS. NWS has 00Z model data available, and has added advantage of watching conditions develop since 00Z. Models of course can’t look at NWS, but NWS looks at models. NWS Forecasts going out 48 (model out 60) hours, so in the analysis there are: Two maximum temperatures (MAX-T), Two minimum temperatures (MIN-T), and Four 12-hr POP forecasts.

Temperature Comparisons

Temperature MAE (F) for the seven forecast types for all stations, all time periods, 1 August 2003 – 1 August 2004.

Precipitation Comparisons

Brier Scores for Precipitation for all stations for the entire study period.

They don’t do traditional MOS!

Dynamic MOS Using Multiple Models MOS equations are updated frequently, not static like the NWS. Multiple model and observation inputs Example: DiCast used by the Weather Channel and Accuweather, Developed by NCAR

Dynamical MOS of MOSs

DICAST skill is quite good

ForecastAdvisor.com

New NWS Approach: National Blend of Models Ghe National Blend of Models (NBM) is a nationally consistent and skillful suite of calibrated forecast guidance based on a blend of both NWS and non-NWS numerical weather prediction model data and post-processed model guidance. The goal of the NBM is to create a highly accurate, skillful and consistent starting point for the gridded forecast.

Blend This first version used 3 models (GFS, GEFS mean, CMCE mean) and provided temperature, wind, and sky cover over the CONUS region two times a day. Added more parameters recently 2.5 km grid

There are many other post-processing approaches Neural nets Attempts to duplicate the complex interactions between neurons in the human brain.

Bayesian Model Averaging (BMA) A good way to optimize the use of ensembles of forecasts to provide calibrated probabilistic guidance. Can weight models and the variability by previous performance.

Machine Learning/Artificial Intelligence