Validating Bayesian Inference Algorithms with Simulation-Based Calibration

18 Apr 2018  ·  Sean Talts, Michael Betancourt, Daniel Simpson, Aki Vehtari, Andrew Gelman ·

Verifying the correctness of Bayesian computation is challenging. This is especially true for complex models that are common in practice, as these require sophisticated model implementations and algorithms. In this paper we introduce \emph{simulation-based calibration} (SBC), a general procedure for validating inferences from Bayesian algorithms capable of generating posterior samples. This procedure not only identifies inaccurate computation and inconsistencies in model implementations but also provides graphical summaries that can indicate the nature of the problems that arise. We argue that SBC is a critical part of a robust Bayesian workflow, as well as being a useful tool for those developing computational algorithms and statistical software.

PDF Abstract

Datasets


  Add Datasets introduced or used in this paper