probability forecasts
Recently Published Documents


TOTAL DOCUMENTS

159
(FIVE YEARS 16)

H-INDEX

27
(FIVE YEARS 2)

2021 ◽  
Vol 9 (12) ◽  
pp. 471-489
Author(s):  
Mary E. Thomson ◽  
Andrew C. Pollock ◽  
Jennifer Murray

An analytical framework is presented for the evaluation of composite probability forecasts using empirical quantiles. The framework is demonstrated via the examination of forecasts of the changes in the number of US COVID-19 confirmed infection cases, applying 18 two-week ahead quantile forecasts from four forecasting organisations. The forecasts are analysed individually for each organisation and in combinations of organisational forecasts to ascertain the highest level of performance. It is shown that the relative error reduction achieved by combining forecasts depends on the extent to which the component forecasts contain independent information. The implications of the study are discussed, suggestions are offered for future research and potential limitations are considered.


2021 ◽  
Author(s):  
David Johnstone ◽  
Stewart Jones ◽  
Oliver Jones ◽  
Steve Tulig

The purpose of our paper is to describe a probability scoring rule that reflects the economic performance of a hypothetical investor who acts upon the probability forecasts emanating from a given model or human expert by trading against a market-clearing consensus of competing models and forecasts. The probability forecasts being compared are aggregated by an equilibrium condition into a market consensus reflecting the wisdom of the crowd. A good forecaster (model or human expert) is measured as one who allows the user to bet profitably against the market consensus. By asking forecasts to beat the market, forecasters are discouraged from herding and motivated to obtain better information than rival forecasters. We illustrate and prove that each trader’s personal incentive to hedge or fudge disappears when the number of forecasts in the market is sufficiently large. Our score exhibits the forecaster’s ability to assist economically profitable action and reveals how the user’s profits depend strongly on the accuracy of the forecasts and the decision rule (boldness) with which they are acted upon.


2021 ◽  
Author(s):  
Robert Mislavsky ◽  
Celia Gaertig

How do we combine others’ probability forecasts? Prior research has shown that when advisors provide numeric probability forecasts, people typically average them (i.e., they move closer to the average advisor’s forecast). However, what if the advisors say that an event is “likely” or “probable?” In eight studies (n = 7,334), we find that people are more likely to act as if they “count” verbal probabilities (i.e., they move closer to certainty than any individual advisor’s forecast) than they are to “count” numeric probabilities. For example, when the advisors both say an event is “likely,” participants will say that it is “very likely.” This effect occurs for both probabilities above and below 50%, for hypothetical scenarios and real events, and when presenting the others’ forecasts simultaneously or sequentially. We also show that this combination strategy carries over to subsequent consumer decisions that rely on advisors’ likelihood judgments. We discuss and rule out several candidate mechanisms for our effect. This paper was accepted by Yuval Rottenstreich, decision analysis.


2021 ◽  
Vol 230 (1) ◽  
pp. 381-407
Author(s):  
Hong-Jia Chen ◽  
Chien-Chih Chen ◽  
Guy Ouillon ◽  
Didier Sornette

2021 ◽  
Vol 9 (2) ◽  
pp. 139-163
Author(s):  
Mary E. Thomson ◽  
◽  
Andrew C. Pollock ◽  
Jennifer Murray ◽  
◽  
...  

An analytical framework is presented for the evaluation of quantile probability forecasts. It is demonstrated using weekly quantile forecasts of changes in the number of US COVID-19 deaths. Empirical quantiles are derived using the assumption that daily changes in a variable follow a normal distribution with time varying means and standard deviations, which can be assumed constant over short horizons such as one week. These empirical quantiles are used to evaluate quantile forecasts using the Mean Squared Quantile Score (MSQS), which, in turn, is decomposed into sub-components involving bias, resolution and error variation to identify specific aspects of performance, which highlight the strengths and weaknesses of forecasts. The framework is then extended to test if performance enhancement can be achieved by combining diverse forecasts from different sources. The demonstration illustrates that the technique can effectively evaluate quantile forecasting performance based on a limited number of data points, which is crucial in emergency situations such as forecasting pandemic behavior. It also shows that combining the predictions with quantile probability forecasts generated from an Autoregressive Order One, AR(1) model provided substantially improved performance. The implications of these findings are discussed, suggestions are offered for future research and potential limitations are considered.


2020 ◽  
Vol 35 (5) ◽  
pp. 1981-2000
Author(s):  
Ryan A. Sobash ◽  
Glen S. Romine ◽  
Craig S. Schwartz

AbstractA feed-forward neural network (NN) was trained to produce gridded probabilistic convective hazard predictions over the contiguous United States. Input fields to the NN included 174 predictors, derived from 38 variables output by 497 convection-allowing model forecasts, with observed severe storm reports used for training and verification. These NN probability forecasts (NNPFs) were compared to surrogate-severe probability forecasts (SSPFs), generated by smoothing a field of surrogate reports derived with updraft helicity (UH). NNPFs and SSPFs were produced each forecast hour on an 80-km grid, with forecasts valid for the occurrence of any severe weather report within 40 or 120 km, and 2 h, of each 80-km grid box. NNPFs were superior to SSPFs, producing statistically significant improvements in forecast reliability and resolution. Additionally, NNPFs retained more large magnitude probabilities (>50%) compared to SSPFs since NNPFs did not use spatial smoothing, improving forecast sharpness. NNPFs were most skillful relative to SSPFs when predicting hazards on larger scales (e.g., 120 vs 40 km) and in situations where using UH was detrimental to forecast skill. These included model spinup, nocturnal periods, and regions and environments where supercells were less common, such as the western and eastern United States and high-shear, low-CAPE regimes. NNPFs trained with fewer predictors were more skillful than SSPFs, but not as skillful as the full-predictor NNPFs, with predictor importance being a function of forecast lead time. Placing NNPF skill in the context of existing baselines is a first step toward integrating machine learning–based forecasts into the operational forecasting process.


2020 ◽  
Vol 96 (314) ◽  
pp. 294-313
Author(s):  
Kevin Lee ◽  
Kian Ong ◽  
Kalvinder K. Shields

Sign in / Sign up

Export Citation Format

Share Document