Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Efficiently Sampling Functions from Gaussian Process Posteriors

About

Gaussian processes are the gold standard for many real-world modeling problems, especially in cases where a model's success hinges upon its ability to faithfully represent predictive uncertainty. These problems typically exist as parts of larger frameworks, wherein quantities of interest are ultimately defined by integrating over posterior distributions. These quantities are frequently intractable, motivating the use of Monte Carlo methods. Despite substantial progress in scaling up Gaussian processes to large training sets, methods for accurately generating draws from their posterior distributions still scale cubically in the number of test locations. We identify a decomposition of Gaussian processes that naturally lends itself to scalable sampling by separating out the prior from the data. Building off of this factorization, we propose an easy-to-use and general-purpose approach for fast posterior sampling, which seamlessly pairs with sparse approximations to afford scalability both during training and at test time. In a series of experiments designed to test competing sampling schemes' statistical properties and practical ramifications, we demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.

James T. Wilson, Viacheslav Borovitskiy, Alexander Terenin, Peter Mostowsky, Marc Peter Deisenroth• 2020

Related benchmarks

TaskDatasetResultRank
High-dimensional optimizationRover 60D
Objective Value-0.33
6
High-dimensional optimizationLassoDNA 180D
Objective Value-0.3
6
High-dimensional optimizationSVM 388D
Objective Value-0.21
6
High-dimensional optimizationMOPTA08 124D
Objective Value-227.8
6
High-dimensional optimizationMedian Molecules 256D 1
Objective Value0.27
5
High-dimensional optimizationMedian Molecules 256D 2
Objective Value0.1
5
Showing 6 of 6 rows

Other info

Follow for update