bayesian aggregation
Recently Published Documents


TOTAL DOCUMENTS

29
(FIVE YEARS 6)

H-INDEX

9
(FIVE YEARS 1)

2021 ◽  
pp. 1-13
Author(s):  
Yuling Yao
Keyword(s):  

Author(s):  
Yakov Babichenko ◽  
Dan Garber

We consider the forecast aggregation problem in repeated settings where the forecasts are of a binary state of nature. In each period multiple experts provide forecasts about the state. The goal of the aggregator is to aggregate those forecasts into a subjective accurate forecast. We assume that the experts are Bayesian and the aggregator is non-Bayesian and ignorant of the information structure (i.e., the distribution over the signals) under which the experts make their forecasts. The aggregator observes the experts’ forecasts only. At the end of each period, the realized state is observed by the aggregator. We focus on the question of whether the aggregator can learn to optimally aggregate the forecasts of the experts, where the optimal aggregation is the Bayesian aggregation that takes into account all the information in the system. We consider the class of partial evidence information structures, where each expert is exposed to a different subset of conditionally independent signals. Our main results are positive: we show that optimal aggregation can be learned in polynomial time in quite a wide range of instances in partial evidence environments. We provide an exact characterization of the instances where optimal learning is possible as well as those where it is impossible.


2019 ◽  
Vol 15 (S341) ◽  
pp. 99-103 ◽  
Author(s):  
Hugh Dickinson ◽  
Lucy Fortson ◽  
Claudia Scarlata ◽  
Melanie Beck ◽  
Mike Walmsley

AbstractLSST and Euclid must address the daunting challenge of analyzing the unprecedented volumes of imaging and spectroscopic data that these next-generation instruments will generate. A promising approach to overcoming this challenge involves rapid, automatic image processing using appropriately trained Deep Learning (DL) algorithms. However, reliable application of DL requires large, accurately labeled samples of training data. Galaxy Zoo Express (GZX) is a recent experiment that simulated using Bayesian inference to dynamically aggregate binary responses provided by citizen scientists via the Zooniverse crowd-sourcing platform in real time. The GZX approach enables collaboration between human and machine classifiers and provides rapidly generated, reliably labeled datasets, thereby enabling online training of accurate machine classifiers. We present selected results from GZX and show how the Bayesian aggregation engine it uses can be extended to efficiently provide object-localization and bounding-box annotations of two-dimensional data with quantified reliability. DL algorithms that are trained using these annotations will facilitate numerous panchromatic data modeling tasks including morphological classification and substructure detection in direct imaging, as well as decontamination and emission line identification for slitless spectroscopy. Effectively combining the speed of modern computational analyses with the human capacity to extrapolate from few examples will be critical if the potential of forthcoming large-scale surveys is to be realized.


2018 ◽  
Vol 12 (3) ◽  
pp. 1583-1604 ◽  
Author(s):  
Sebastian Weber ◽  
Andrew Gelman ◽  
Daniel Lee ◽  
Michael Betancourt ◽  
Aki Vehtari ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document