Flexible Dataset Distillation: Learn Labels Instead of Images

Research output: Contribution to conferencePaperpeer-review

Abstract / Description of output

We study the problem of dataset distillation - creating a small set of synthetic examples capable of training a good model. In particular, we study the problem of label distillation - creating synthetic labels for a small set of real images, and show it to be more effective than the prior image-based approach to dataset distillation. Methodologically, we introduce a more robust and flexible meta-learning algorithm for distillation, as well as an effective first-order strategy based on convex optimization layers. Distilling labels with our new algorithm leads to improved results over prior image-based distillation. More importantly, it leads to clear improvements in flexibility of the distilled dataset in terms of compatibility with off-the-shelf optimizers and diverse neural architectures. Interestingly, label distillation can also be applied across datasets, for example enabling learning Japanese character recognition by training only on synthetically labeled English letters.
Original languageEnglish
Pages1-21
DOIs
Publication statusPublished - 12 Dec 2020
Event4th Workshop on Meta-Learning at NeurIPS 2020 - Vancouver, Canada
Duration: 11 Dec 2020 → …
Conference number: 4
https://nips.cc/virtual/2020/public/e_workshops.html

Workshop

Workshop4th Workshop on Meta-Learning at NeurIPS 2020
Country/TerritoryCanada
CityVancouver
Period11/12/20 → …
Internet address

Fingerprint

Dive into the research topics of 'Flexible Dataset Distillation: Learn Labels Instead of Images'. Together they form a unique fingerprint.

Cite this