Abstract
An efficient learner is one who reuses what they already know to tackle a new problem. For a machine learner, this means understanding the similarities amongst datasets. In order to do this, one must take seriously the idea of working with datasets, rather than datapoints, as the key objects to model. Towards this goal, we demonstrate an extension of a variational autoencoder that can learn a method for computing representations, or statistics, of datasets in an unsupervised fashion. The network is trained to produce statistics that encapsulate a generative model for each dataset. Hence the network enables efficient learning from new datasets for both unsupervised and supervised tasks. We show that we are able to learn statistics that can be used for: clustering datasets, transferring generative models to new datasets, selecting representative samples of datasets and classifying previously unseen classes. We refer to our model as a neural statistician, and by this we mean a neural network that can learn to compute summary statistics of datasets without supervision.
Original language | English |
---|---|
Title of host publication | 5th International Conference on Learning Representations (ICLR 2017) |
Pages | 1-13 |
Number of pages | 13 |
Publication status | Published - 26 Apr 2017 |
Event | 5th International Conference on Learning Representations - Palais des Congrès Neptune, Toulon, France Duration: 24 Apr 2017 → 26 Apr 2017 https://iclr.cc/archive/www/2017.html |
Conference
Conference | 5th International Conference on Learning Representations |
---|---|
Abbreviated title | ICLR 2017 |
Country/Territory | France |
City | Toulon |
Period | 24/04/17 → 26/04/17 |
Internet address |
Fingerprint
Dive into the research topics of 'Towards a Neural Statistician'. Together they form a unique fingerprint.Profiles
-
Amos Storkey
- School of Informatics - Personal Chair of Machine Learning & Artificial Intelligence
- Institute for Adaptive and Neural Computation
- Data Science and Artificial Intelligence
Person: Academic: Research Active