1 / 24

Probabilistic Forecasting

Probabilistic Forecasting. pdfs and Histograms. A few different normal (Gaussian) pdfs. Probability density functions (pdfs) are unobservable. They can only be estimated. They tell us the density, and must be integrated to get the probability. pdfs and Histograms.

loe
Download Presentation

Probabilistic Forecasting

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Probabilistic Forecasting

  2. pdfs and Histograms A few different normal (Gaussian) pdfs • Probability density functions (pdfs) are unobservable. They can only be estimated. • They tell us the density, and must be integrated to get the probability.

  3. pdfs and Histograms • Histograms are already integrated over the chosen bin width, and provide an estimated probability. • One might fit a function to a histogram to arrive at a pdf.

  4. pdfs and Histograms • Probability density functions (pdfs) are unobservable. They can only be estimated. • They tell us the density, and must be integrated to get the probability.

  5. cdfs and thresholds A few different normal (Gaussian) cdfs • Can integrate from from one point to infinity to get the cumulative distribution function (cdf)

  6. cdfs and thresholds • Histograms can also be accumulated. • One might fit a function to a cumulative histogram to arrive at a cdf.

  7. pdfs and cdfs

  8. Verifying probabilistic forecasts for usefulness • Reliability: agreement between forecast frequency/probability and observed frequency • Resolution: ability of a forecast to discriminate between events • Sharpness: tendency to forecast event probabilities of 0 or 1 instead of clustering around the mean

  9. Complementary metrics • Forecast conditioned on the observations • Observations conditioned on the forecasts

  10. Reliability • Rank Histogram: How well does the ensemble spread in the forecast represent uncertainty, on average? • Reliability Diagram: How well do the predicted probabilities of an event correspond to their observed frequencies?

  11. Rank Histogram • U-shaped: observations usually outside of ensemble envelope; underdispersive ensemble • Flat: observations usually indistinguishable from the members of the ensemble • Humped: observations usually in the middle of the ensemble; overdispersive ensemble

  12. Reliability Diagram • Given that X was predicted with probability Y, what was the outcome? • How well do the observations of an event correspond to the predicted probabilities? • A forecast of climatology has no reolution.

  13. Resolution • Given that X was observed with probability Y, what was the forecast? • How well did the probability forecast predict the category bin containing the observation?

  14. Calibration • Probabilistic calibration is necessary because the model cannot produce the observed distribution • This includes correcting both the bias (mean) and the variability (spread)

  15. Calibration

  16. Test Environment • Lot-acceptance ammunition testing • Planning and test completion thresholds of 5 and 7 m/s crosswinds • Peak winds (gust) on-site decisions

  17. Probabilistic Forecasts for Direct-Fire Ballistics An ensemble of wind forecasts Crosswind component: Calibrated forecast distribution Firing Range

  18. Probability Forecasts for Direct-Fire Ballistics Time series: probability of exceedance 5 ms-1 threshold 7 ms-1 threshold Cumulative distribution for a given time

  19. Goal: Reliable Ensembles for Crosswind Thresholds • Over several forecasts, the verification is statistically indistinguishable from the ensemble. • Model error must be taken into account (calibration). • Reliability is the first step, later we will consider resolution.

  20. Calibration We are shooting for this from the model: These distributions are lognormal, and we correct the mean and variance in the same way.

  21. Forecast vs. Observed • Forecast has a large positive bias in wind speed • False positive forecasts for winds > 5 m/s 28% of the time. False Negative Observed wind speed (m/s) False Positive Forecast wind speed (m/s)

  22. Simple Solutions Inadequate • Linear regression to correct • Removes false positives • Introduces more false negatives • Bimodality may be a problem False Negative Observed wind speed (m/s) False Positive Adjusted forecast wind speed (m/s)

  23. Monthly Variability • Distributions of regression residuals each month • Shows that a single calibration for all times is not appropriate Linear Model Residuals Month

  24. Summary • We want an to estimate pdf useful for decision making (gambling). • An ensemble forecast can be the basis. • Calibration is necessary, but can be difficult.

More Related