Variational Gibbs inference for statistical model estimation from incomplete data

Vaidotas Simkus, Benjamin Rhodes, Michael U. Gutmann

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

Statistical models are central to machine learning with broad applicability across a range of downstream tasks. The models are controlled by free parameters that are typically estimated from data by maximum-likelihood estimation or approximations thereof. However, when faced with real-world datasets many of the models run into a critical issue: they are formulated in terms of fully-observed data, whereas in practice the datasets are plagued with missing data. The theory of statistical model estimation from incomplete data is conceptually similar to the estimation of latent-variable models, where powerful tools such as variational inference (VI) exist. However, in contrast to standard latent-variable models, parameter estimation with incomplete data often requires estimating exponentially-many conditional distributions of the missing variables, hence making standard VI methods intractable. We address this gap by introducing variational Gibbs inference (VGI), a new general-purpose method to estimate the parameters of statistical models from incomplete data. We validate VGI on a set of synthetic and real-world estimation tasks, estimating important machine learning models such as VAEs and normalising flows from incomplete data. The proposed method, whilst general-purpose, achieves competitive or better performance than existing model-specific estimation methods.
Original languageEnglish
Article number196
Pages (from-to)1-72
Number of pages72
JournalJournal of Machine Learning Research
Volume24
Publication statusPublished - 30 Jun 2023

Fingerprint

Dive into the research topics of 'Variational Gibbs inference for statistical model estimation from incomplete data'. Together they form a unique fingerprint.

Cite this