An IRT forecasting model: linking proper scoring rules to item response theory

JUDGMENT AND DECISION MAKING(2017)

引用 30|浏览10
暂无评分
摘要
This article proposes an Item Response Theoretical (IRT) forecasting model that incorporates proper scoring rules and provides evaluations of forecasters' expertise in relation to the features of the specific questions they answer. We illustrate the model using geopolitical forecasts obtained by the Good Judgment Project (GJP) (see Mellers, Ungar, Baron, Ramos, Gurcay, Fincher, Scott, Moore, Atanasov, Swift, Murray, Stone & Tetlock, 2014). The expertise estimates from the IRT model, which take into account variation in the difficulty and discrimination power of the events, capture the underlying construct being measured and are highly correlated with the forecasters' Brier scores. Furthermore, our expertise estimates based on the first three years of the GJP data are better predictors of both the forecasters' fourth year Brier scores and their activity level than the overall Brier scores obtained and Merkle's (2016) predictions, based on the same period. Lastly, we discuss the benefits of using event-characteristic information in forecasting.
更多
查看译文
关键词
IRT,Forecasting,Brier scores,Proper Scoring Rules,Good Judgment Project,Gibbs sampling
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要