Projects per year
Abstract
Most contemporary multi-task learning methods assume linear models. This setting is considered shallow in the era of deep learning. In this paper, we present a new deep multi-task representation learning framework that learns cross-task sharing structure at every layer in a deep network. Our approach is based on generalising the matrix factorisation techniques explicitly or implicitly used by many conventional MTL algorithms to tensor factorisation, to realise automatic learning of end-to-end knowledge sharing in deep networks. This is in contrast to existing deep learning approaches that need a user-defined multi-task sharing strategy. Our approach applies to both homogeneous and heterogeneous MTL. Experiments demonstrate the efficacy of our deep multi-task representation learning in terms of both higher accuracy and fewer design choices.
Original language | English |
---|---|
Title of host publication | International Conference on Learning Representations (ICLR 2017) |
Number of pages | 12 |
Publication status | E-pub ahead of print - 26 Apr 2017 |
Event | 5th International Conference on Learning Representations - Palais des Congrès Neptune, Toulon, France Duration: 24 Apr 2017 → 26 Apr 2017 https://iclr.cc/archive/www/2017.html |
Conference
Conference | 5th International Conference on Learning Representations |
---|---|
Abbreviated title | ICLR 2017 |
Country/Territory | France |
City | Toulon |
Period | 24/04/17 → 26/04/17 |
Internet address |
Fingerprint
Dive into the research topics of 'Deep Multi-task Representation Learning: A Tensor Factorisation Approach'. Together they form a unique fingerprint.Projects
- 1 Finished
-
DREAM - Deferred Restructuring of Experience in Autonomous Machines
1/09/16 → 31/12/18
Project: Research
Profiles
-
Timothy Hospedales
- School of Informatics - Personal Chair of Artificial Intelligence
- Institute of Perception, Action and Behaviour
- Language, Interaction and Robotics
Person: Academic: Research Active