Next Article in Journal
Statistical Inference on the Shannon Entropy of Inverse Weibull Distribution under the Progressive First-Failure Censoring
Previous Article in Journal
Towards Generation of Cat States in Trapped Ions Set-Ups via FAQUAD Protocols and Dynamical Decoupling
Open AccessArticle

Learning from Both Experts and Data

CMAP Ecole Polytechnique, Institut Polytechnique de Paris, 91128 Palaiseau, France
XPop, Inria Saclay, 91120 Palaiseau, France
School of Medecine, Université Paris-Descartes, 75006 Paris, France
Author to whom correspondence should be addressed.
Entropy 2019, 21(12), 1208;
Received: 29 October 2019 / Revised: 29 November 2019 / Accepted: 6 December 2019 / Published: 10 December 2019
(This article belongs to the Section Information Theory, Probability and Statistics)
In this work, we study the problem of inferring a discrete probability distribution using both expert knowledge and empirical data. This is an important issue for many applications where the scarcity of data prevents a purely empirical approach. In this context, it is common to rely first on an a priori from initial domain knowledge before proceeding to an online data acquisition. We are particularly interested in the intermediate regime, where we do not have enough data to do without the initial a priori of the experts, but enough to correct it if necessary. We present here a novel way to tackle this issue, with a method providing an objective way to choose the weight to be given to experts compared to data. We show, both empirically and theoretically, that our proposed estimator is always more efficient than the best of the two models (expert or data) within a constant. View Full-Text
Keywords: maximum entropy; mixing expert and data; Kullback–Leibler centroid maximum entropy; mixing expert and data; Kullback–Leibler centroid
Show Figures

Figure 1

MDPI and ACS Style

Besson, R.; Le Pennec, E.; Allassonnière, S. Learning from Both Experts and Data. Entropy 2019, 21, 1208.

AMA Style

Besson R, Le Pennec E, Allassonnière S. Learning from Both Experts and Data. Entropy. 2019; 21(12):1208.

Chicago/Turabian Style

Besson, Rémi; Le Pennec, Erwan; Allassonnière, Stéphanie. 2019. "Learning from Both Experts and Data" Entropy 21, no. 12: 1208.

Find Other Styles
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Article Access Map by Country/Region

Search more from Scilit
Back to TopTop