Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Sequential Neural Likelihood: Fast Likelihood-free Inference with Autoregressive Flows

About

We present Sequential Neural Likelihood (SNL), a new method for Bayesian inference in simulator models, where the likelihood is intractable but simulating data from the model is possible. SNL trains an autoregressive flow on simulated data in order to learn a model of the likelihood in the region of high posterior density. A sequential training procedure guides simulations and reduces simulation cost by orders of magnitude. We show that SNL is more robust, more accurate and requires less tuning than related neural-based methods, and we discuss diagnostics for assessing calibration, convergence and goodness-of-fit.

George Papamakarios, David C. Sterratt, Iain Murray• 2018

Related benchmarks

TaskDatasetResultRank
Source FindingSource Finding (2D)
Avg sPCE Lower Bound9.62
9
Source FindingSource Finding 3D
Avg sPCE Lower Bound5.37
9
Source FindingSource Finding 5D
Average sPCE LB3.01
8
Simulation-Based Inferenceg-and-k distribution Huber contamination (90% clean, 10% outliers) (test)
MMD2 (Ref)0.54
8
Simulation-Based InferenceRadio propagation simulator
Inference Time (s)13.7
4
Showing 5 of 5 rows

Other info

Follow for update