Trace-class Gaussian priors for Bayesian learning of neural networks with MCMC

Torben Sell, Sumeetpal Sidhu Singh

Research output: Contribution to journalArticlepeer-review

Abstract

This paper introduces a new neural network based prior for real valued functions. Each weight and bias of the neural network has an independent Gaussian prior, with the key novelty that the variances decrease in the width of the network in such a way that the resulting function is well defined in the limit of an infinite width network. We show that the induced posterior over functions is amenable to Monte Carlo sampling using Hilbert space Markov chain Monte Carlo (MCMC) methods. This type of MCMC is stable under mesh refinement, i.e. the acceptance probability does not degenerate as more parameters of the function's prior are introduced, even ad infinitum. We demonstrate these advantages over other function space priors, for example in Bayesian Reinforcement Learning.
Original languageEnglish
Pages (from-to)46-66
Number of pages21
JournalJournal of the Royal Statistical Society: Statistical Methodology Series B
Volume85
Issue number1
DOIs
Publication statusPublished - 31 Jan 2023

Fingerprint

Dive into the research topics of 'Trace-class Gaussian priors for Bayesian learning of neural networks with MCMC'. Together they form a unique fingerprint.

Cite this