Universal representations: A unified look at multiple task and domain learning

Wei-Hong Li*, Xialei Liu, Hakan Bilen

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

We propose a unified look at jointly learning multiple vision tasks and visual domains through universal representations, a single deep neural network. Learning multiple problems simultaneously involves minimizing a weighted sum of multiple loss functions with different magnitudes and characteristics and thus results in unbalanced state of one loss dominating the optimization and poor results compared to learning a separate model for each problem. To this end, we propose distilling knowledge of multiple task/domain-specific networks into a single deep neural network after aligning its representations with the task/domain-specific ones through small capacity adapters. We rigorously show that universal representations achieve state-of-the-art performances in learning of multiple dense prediction problems in NYU-v2 and Cityscapes, multiple image classification problems from diverse domains in Visual Decathlon Dataset and cross-domain few-shot learning in MetaDataset. Finally we also conduct multiple analysis through ablation and qualitative studies.
Original languageEnglish
Pages (from-to)1521–1545
Number of pages25
JournalInternational Journal of Computer Vision
Volume132
DOIs
Publication statusPublished - 24 Nov 2023

Keywords / Materials (for Non-textual outputs)

  • multi-task learning
  • multi-domain learning
  • cross-domain few-shot learning
  • universal representation learning
  • balanced optimization
  • dense prediction

Fingerprint

Dive into the research topics of 'Universal representations: A unified look at multiple task and domain learning'. Together they form a unique fingerprint.

Cite this