News & Events

Subscribe to email list

Please select the email list(s) to which you wish to subscribe.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Image CAPTCHA

Enter the characters shown in the image.

User menu

You are here

Automated, Scalable Bayesian Inference with Theoretical Guarantees

Tuesday, February 27, 2018 - 11:00 to 12:00
Trevor Campbell, Post-doctoral Associate at MIT
Statistics Seminar
Room 4192, Earth Sciences Building (2207 Main Mall)

The automation of posterior inference in Bayesian data analysis has enabled experts and nonexperts alike to use more sophisticated models, engage in faster exploratory modeling and analysis, and ensure experimental reproducibility. However, standard automated posterior inference algorithms are not tractable at the scale of massive modern datasets, and modifications to make them so are typically model-specific, require expert tuning, and can break theoretical guarantees on inferential quality. This talk will instead take advantage of data redundancy to shrink the dataset itself as a preprocessing step, forming a "Bayesian coreset." The coreset can be used in a standard inference algorithm at significantly reduced cost while maintaining theoretical guarantees on posterior approximation quality. The talk will include an intuitive formulation of Bayesian coreset construction as sparse vector sum approximation, an automated coreset construction algorithm that takes advantage of this formulation, strong theoretical guarantees on posterior approximation quality, and applications to a variety of real and simulated datasets.