Learn Prior distributions from expert knowledge#

Note

This website is currently under construction.

Please let me know if you notice any mistakes or omissions.

Short overview#

We propose a new elicitation method for translating knowledge from a domain expert into an appropriate parametric prior distribution (Bockting et al., 2023). In particular, our approach builds on recent contributions made by Hartmann et al. (2020), da Silva et al. (2019), and Manderson & Goudie (2023). Their key commonality is the development of (more or less) model-agnostic methods in which the search for appropriate prior distributions is formulated as an optimization problem. Thus, the objective is to determine the optimal hyperparameters that minimize the discrepancy between model-implied and expert-elicited statistics. We also adopt this perspective and introduce a novel elicitation method that supports expert feedback in both the space of parameters and observable quantities (i.e., a hybrid approach) and minimizes human effort. The key ideas underlying our method are visualized in the following figure and outlined below.

Visual representation workflow
  1. The analyst defines a generative model comprising a likelihood function \(p(y \mid \theta)\) and a parametric prior distribution \(p(\theta \mid \lambda)\) for the model parameters, where \(\lambda\) represents the prior hyperparameters to be inferred from expert knowledge.

  2. The analyst selects a set of target quantities, which may involve queries related to observable quantities (data), model parameters, or anything else in between.

  3. The domain expert is queried using a specific elicitation technique for each target quantity (expert-elicited statistics).

  4. From the generative model implied by likelihood and prior and a given value of \(\lambda\), parameters and (prior) predictive data are simulated, and the predefined set of target quantities is computed based on the simulations (model-implied quantities).

  5. The discrepancy between the model-implied and the expert-elicited statistics is evaluated with a discrepancy measure (loss function).

  6. Stochastic gradient descent is employed to update the hyperparameters \(\lambda\) so as to minimize the loss function.

  7. Steps 4 to 6 are repeated iteratively until an optimal set of hyperparameters \(\lambda\) is found that minimizes the discrepancy between the model-implied and the expert-elicited statistics.

References#

Contents#