Towards a Neural Statistician

Harrison Edwards, Amos Storkey

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

An efficient learner is one who reuses what they already know to tackle a new problem. For a machine learner, this means understanding the similarities amongst datasets. In order to do this, one must take seriously the idea of working with datasets, rather than datapoints, as the key objects to model. Towards this goal, we demonstrate an extension of a variational autoencoder that can learn a method for computing representations, or statistics, of datasets in an unsupervised fashion. The network is trained to produce statistics that encapsulate a generative model for each dataset. Hence the network enables efficient learning from new datasets for both unsupervised and supervised tasks. We show that we are able to learn statistics that can be used for: clustering datasets, transferring generative models to new datasets, selecting representative samples of datasets and classifying previously unseen classes. We refer to our model as a neural statistician, and by this we mean a neural network that can learn to compute summary statistics of datasets without supervision.
Original languageEnglish
Title of host publication5th International Conference on Learning Representations (ICLR 2017)
Pages1-13
Number of pages13
Publication statusPublished - 26 Apr 2017
Event5th International Conference on Learning Representations - Palais des Congrès Neptune, Toulon, France
Duration: 24 Apr 201726 Apr 2017
https://iclr.cc/archive/www/2017.html

Conference

Conference5th International Conference on Learning Representations
Abbreviated titleICLR 2017
Country/TerritoryFrance
CityToulon
Period24/04/1726/04/17
Internet address

Fingerprint

Dive into the research topics of 'Towards a Neural Statistician'. Together they form a unique fingerprint.

Cite this