Subjective Verification of Numerical Models as a Component of a Broader Interaction between Research and Operations
- 1 October 2003
- journal article
- Published by American Meteorological Society in Weather and Forecasting
- Vol. 18 (5), 847-860
- https://doi.org/10.1175/1520-0434(2003)018<0847:svonma>2.0.co;2
Abstract
Systematic subjective verification of precipitation forecasts from two numerical models is presented and discussed. The subjective verification effort was carried out as part of the 2001 Spring Program, a seven-week collaborative experiment conducted at the NOAA/National Severe Storms Laboratory (NSSL) and the NWS/Storm Prediction Center, with participation from the NCEP/Environmental Modeling Center, the NOAA/Forecast Systems Laboratory, the Norman, Oklahoma, National Weather Service Forecast Office, and Iowa State University. This paper focuses on a comparison of the operational Eta Model and an experimental version of this model run at NSSL; results are limited to precipitation forecasts, although other models and model output fields were verified and evaluated during the program. By comparing forecaster confidence in model solutions to next-day assessments of model performance, this study yields unique information about the utility of models for human forecasters. It is shown that, when averaged over many forecasts, subjective verification ratings of model performance were consistent with preevent confidence levels. In particular, models that earned higher average confidence ratings were also assigned higher average subjective verification scores. However, confidence and verification scores for individual forecasts were very poorly correlated, that is, forecast teams showed little skill in assessing how “good” individual model forecasts would be. Furthermore, the teams were unable to choose reliably which model, or which initialization of the same model, would produce the “best” forecast for a given period. The subjective verification methodology used in the 2001 Spring Program is presented as a prototype for more refined and focused subjective verification efforts in the future. The results demonstrate that this approach can provide valuable insight into how forecasters use numerical models. It has great potential as a complement to objective verification scores and can have a significant positive impact on model development strategies.Keywords
This publication has 16 references indexed in Scilit:
- Comparative Verification of Recent Quantitative Precipitation Forecasts in the National Weather Service: A Simple Approach for Scoring Forecast AccuracyWeather and Forecasting, 2003
- Properties of the Convection Scheme in NCEP's Eta Model that Affect Forecast Sounding InterpretationWeather and Forecasting, 2002
- Comparison between Observed Convective Cloud-Base Heights and Lifting Condensation Level for Two Different Lifted ParcelsWeather and Forecasting, 2002
- Examination of Derecho Environments Using Proximity SoundingsWeather and Forecasting, 2001
- The Melting Effect as a Factor in Precipitation-Type ForecastingWeather and Forecasting, 2000
- Hypothesis Tests for Evaluating Numerical Precipitation ForecastsWeather and Forecasting, 1999
- The New NMC Mesoscale Eta Model: Description and Forecast ExamplesWeather and Forecasting, 1994
- The Step-Mountain Eta Coordinate Model: Further Developments of the Convection, Viscous Sublayer, and Turbulence Closure SchemesMonthly Weather Review, 1994
- Severe Local Storms ForecastingWeather and Forecasting, 1992
- Forecasting and Verifying in a Field Research Project: DOPLIGHT '87Weather and Forecasting, 1989