1 / 34

National Weather Service River Forecast Verification

National Weather Service River Forecast Verification. Peter Gabrielsen July 2006 Hic Meeting July 10, 2006. Background.

tobit
Download Presentation

National Weather Service River Forecast Verification

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. National Weather Service River Forecast Verification Peter Gabrielsen July 2006 Hic Meeting July 10, 2006

  2. Background • As a result of the 2005 NOAA Audit Plan – “The Assistant Administrator for Weather Services should develop, document and implement a timeline and action plan for completing a comprehensive river forecast verification program as soon as practicable” • In 1996 the NRC stated the verification of hydrologic forecasts are inadequate

  3. Background (cont.) • Research has shown: • Little is known about the skill of hydrologic forecasts • Forecasts depend upon imperfect, mathematical descriptions governing runoff and routing • Hydrologic forecasts depend on meteorological forecasts, therefore, they include the uncertainty of meteorological forecasts • Verification leads to improved forecast skill

  4. Background (cont.) • Team was chartered November 2005 • Representatives from five NWS regions and OHD • Expert input from trusted scientists are being used • OHD • OCWWS • Universities • RFCs

  5. Team Charter • Vision: Provide easy access to enhanced river forecast verification data which will be used to improve our scientific and operational techniques and services. • Mission: Assess forecaster, program management and user needs for verification data. Inventory current national and regional verification practices and identify unmet needs. Establish requirements for a comprehensive national system to verify hydrologic forecasts and guidance products which satisfy these needs. This system should identify sources of error and skill in the forecasts across the entire forecast process.

  6. Charter (cont.) • Success Criteria/Deliverables: Deliver a NWS river forecast verification plan which measures skill and error in the forecast process. The plan includes conceptualized solution and a definition of operational requirements

  7. Charter (cont.) • Team Membership: • Julie Demargne (OHD) • Peter Gabrielsen (ER) • Bill Lawrence (SR) • Scott Lindsey (AR) • Mary Mullusky (OCWWS) • Noreen Schwein (CR) • Scott Staggs (WR) • Kevin Werner (WR) • Tom Adams (ER) • William Marosi (NWSEO)

  8. Verification System • Prior to proposing verification standards – the hydrologic forecast process must be described

  9. RFC Hydrologic Forecasting Process Model Calibration Historical data analysis Parameter Estimation Parameter Calibration Operational Implementation Model Parameters Data Processing and Quality Control ObservedForecast Precipitation Precipitation Temperature Temperature Stage Stage Flow Flow Snow depth Freezing level Dewpoint Wind speed Sky Cover Freezing Level Snow water equivalent Potential Evaporation • Hydrologic and Hydraulic Models • Rainfall/Runoff • Snow accumulation …and ablation • Unit Graph • Consumptive Use • Routing • Dynamic Routing • Rating Curves • Reservoir • Statistical Water …Supply (SWS) Postprocessor Short-term Deterministic Forecast Final Product Issuance comparison to action-required stage; appropriate action pursued Long-term Probabilistic Forecast (Ensemble) Short-term Probabilistic Forecast (Ensemble) Forecaster Analysis Review Quality Control Run-time mods Reality Check Long-term Statistical Forecast (Water Supply) Data Assimilation Model States

  10. Input errors and model errors (parameters, model states, model structure) Raw Model Hydrologic Forecast Contribution of RFC staff correcting bad data Data QC to correct input errors Contribution of hydrologic forecaster through runtime-mods Runtime-mods to correct input and model errors (parameters, model states) Adjustments of observed and forecast data (QPF/MPE, MAT, etc.) Contribution of HAS function Operational Hydrologic Forecast Forecast Error Enhanced calibration to correct model parameter errors Enhanced data assimilation process to correct initial model states errors Contribution of forecast processing enhancements Enhanced post-processor to correct output forecast errors Enhanced/new input data to correct input and model errors Experimental / Operational Hydrologic Forecast Enhanced/new hydrologic/hydraulic model to correct model deficiencies Perfect Hydrologic Forecast Corrections of all input, model, and forecaster analysis errors

  11. Role and Setup of the Verification System • Purpose • Monitor forecast quality over time • Monitor quality at various steps in the forecast process • Improves forecast quality • Assist prioritization of forecast system enhancements

  12. Uses of Verification Results • Verification System • Describe forecast performance • Past and recent • Operational • Control (or baseline) • Experimental • Specific time periods

  13. Model Setup: Calibration Operation Installation State Updating: Data Quality Control Runtime Simulations Data Assimilation Verification Forecast Computation: Hydrologic and Hydraulic Models Postprocessor Product Review and Issuance: Forecaster Analysis and Quality Control An effective verification process must quantify the characteristics of the forecast system and offer a means to analyze why forecasts behave the way they do at various steps in the forecast process

  14. Uses of Verification Results • Customers • Hydrologic program managers • Emergency managers • Scientists/Researchers • Hydrologic forecasters • Everyday customers • Use Modes • Operational • Experimental/Research

  15. Verification System Components Administrative – describe the efficiency Logistical aspects – type, quantity, duration and frequency Forecast skill Scientific – describe the reliability Forecast skill Forecast system error analysis

  16. National Baseline Verification System • Logistical • characterizing point forecasts by service type, frequency and location; • characterizing areal forecasts by service type, frequency and location; • identifying daily the number of issued forecasts by type and location; • quantifying the person effort required to set up a basin for forecasting, including data gathering, calibration, model setup and implementation efforts; • quantifying the person effort required to issue each type of forecast, including manual quality control of input data, forecaster run-time modifications and forecaster review and analysis; • quantifying the timeliness of issued forecasts

  17. Categories of Verification Metrics • Categorical: statistics related to predefined threshold or range of values (e.g., above flood stage, minor). • Error: statistics that measure various differences between forecast and observed values (including timing errors). • Correlation: statistics that measure the correspondence between ordered pairs (e.g., crest forecasts vs. QPF, forecast and observed stages). • Distribution Properties:statistics that summarize the characteristics of a set of values.

  18. Categories of Verification Metrics • Skill Scores:statistics that measure the relative accuracy with respect to some set of standard reference or control set of forecasts. • Conditional Statistics:metrics computed based on the occurrence of a particular event or events such as a specific range of observations or forecasts. • Statistical Significance:measures the uncertainty of the computed values of verification metrics.

  19. Verification Systems • National Baseline Verification System • Administrative in nature • Logistical measures • Skill measures • Comprehensive Verification System • Administrative • Scientific

  20. Verification System Requirements • Selection of forecasts to be verified • time attributes (days, months, seasons, years, as well as lead time) • service attributes (national, regional, RFCs, groups, locations) • individual forecaster within guidelines agreed to by the NWS and the NWSEO • basin attributes (response time, size, slope, aspect, elevation, snow, non-snow) • forecast or observed events (crest timing, rising and falling hydrographs)

  21. Verification System Requirements • Archiving • Time attributes (days, months, years, seasons) • Service attributes (national, regional, RFCs, forecaster, groups, locations) • Basin attributes (response time, size, slope, aspect, elevation • Hindcasting • Different QPFs (e.g., Perfect QPF, zero, actual, persistence) • Different FMATs (e.g., Perfect FMAT, actual, persistence) • Different freezing levels • Different MAPEs • Different reservoirs forecasts • Different QPEs (e.g., point based MAP, MAPX, Q2) • Different sets of model parameters • Different models, including the post-processing and state updating models

  22. Additional recommendations • OHD should assign a program manager for verification. • Establish formal verification focal points at each RFC. • Create national river forecast performance goals. This should be accomplished once the software has been fielded and some experience gained with the metrics. • Ensure adequate hydrologic verification training, and use of the system, is captured in OSIP documentation. • Publish findings in peer reviewed journals (e.g., BAMS, EOS) to inform the research community of our plans. • Ensure an end-to-end assessment and verification of the elements in the hydrologic forecasting process that are outside of the control of the RFC forecaster or produced by other agencies

  23. Additional recommendations • OHD needs to establish a team to define the raw model to enable the users to assess the impact of various steps (e.g., calibration, quality control, run-time modifications) on the forecast performance. • Archive of necessary data to support verification software should begin within 30 days of the data being defined. • Ensure continuity with other activities that support this verification plan. • Brief the National Performance Management Committee (NPMC) and ensure incorporation of the RFC hydrologic verification requirements

  24. Background Information

  25. National Baseline Verification System Metrics Categorical Deterministic: POD, FAR, LTD Probabilistic: Brier Score, Ranked Probability Score Error (Accuracy) Deterministic: RMSE, MAE, ME, Bias Correlation: Deterministic: Pearson Correlation Coefficient

  26. National Baseline Verification System Metrics Skill Score Deterministic: RMSE Skill Score Probabilistic: Rank Probability Skill Score, Brier Skill Score Confidence Deterministic: Sample size Probabilistic: Sample size Probabilistic forecasts should also be verified as deterministic forecast using mean or some predetermined exceedence level

  27. Verification System Requirements • Analysis of skill and error sources • Impact of input data errors • Impact of model errors • Impact of forecast analysis • Computation of verification metrics and results presentation • Dissemination and training

  28. Verification metric categories and metrics for deterministic and probabilistic forecasts

  29. Verification metric categories and metrics for deterministic and probabilistic forecasts

  30. Definition of Metrics for the National Baseline Verification System • Probability of detection (POD)– Percentage of (categorical) events forecast correctly. • False Alarm Ration (FAR)– Percentage of (categorical) forecast events that did not verify. • Lead Time of Detection (LTD) – The average lead time of all forecasts that fall into the correct observed category.

  31. Definition of Metrics for the National Baseline Verification System • Root Mean Square Error (RMSE)– The square root of the average of the squared differences between forecasts and observations. • Mean Absolute Error (MAE)– The average of the absolute value of the differences between forecasts and observations. • Mean Error (ME) – The average difference between forecasts and observations. • Bias (%) – The ME expressed as a percentage of the mean observation.

  32. Definition of Metrics for the National Baseline Verification System • Brier Skill Score (BSS) – A skill score based on BS values. The recommended reference forecasts are persistence and climatology. • Ranked Probability Skill Score (RPSS) – A skill score based on RPS values. The recommended reference forecasts are persistence and climatology. • Sample Size – A numeration of the number of forecasts involved in the calculation of a metric appropriate to the type of forecast (e.g., categorical forecasts should numerate forecasts and observations by categories, etc.)

  33. Definition of Metrics for the National Baseline Verification System • Brier Score (BS) - The mean squared error of probabilistic two-category forecasts where the observations are either 0 (no occurrence) or 1 (occurrence) and forecast probability may be arbitrarily distributed between occurrence and non-occurrence. • Ranked Probability Score (RPS) – The mean squared error of probabilistic multi-category forecasts where observations are 1 (occurrence) for the observed category and 0 for all other categories and forecast probability may be arbitrarily distributed between all categories.

  34. Definition of Metrics for the National Baseline Verification System • Correlation Coefficient– A measure of the linear association between forecasts and observations. • Skill Score –In general, skill scores are the percentage difference between verification scores for two sets of forecasts (e.g., operational forecasts and climatology). • Root Mean Squared Error Skill Score (SS-RMSE) – A skill score based on RMSE values. The recommended reference forecasts are persistence and climatology.

More Related