Projects per year
Abstract / Description of output
The transformer is a state-of-the-art neural translation model that uses attention to iteratively refine lexical representations with information drawn from the surrounding context. Lexical features are fed into the first layer and propagated through a deep network of hidden layers. We argue that the need to represent and propagate lexical features in each layer limits the model’s capacity for learning and representing other information relevant to the task. To alleviate this bottleneck, we introduce gated shortcut connections between the embedding layer and each subsequent layer within the encoder and decoder.
This enables the model to access relevant lexical content dynamically, without expending limited resources on storing it within intermediate states. We show that the proposed modification yields consistent improvements over a baseline transformer on standard WMT translation tasks in 5 translation directions (0.9 BLEU on average) and reduces the amount of lexical information passed along the hidden layers. We furthermore evaluate different ways to integrate lexical connections into the transformer architecture and present ablation experiments exploring the effect of proposed shortcuts on model behavior.1
1Our code is publicly available to aid the reproduction of the reported results: https://github.com/demelin/ transformer_lexical_shortcuts
This enables the model to access relevant lexical content dynamically, without expending limited resources on storing it within intermediate states. We show that the proposed modification yields consistent improvements over a baseline transformer on standard WMT translation tasks in 5 translation directions (0.9 BLEU on average) and reduces the amount of lexical information passed along the hidden layers. We furthermore evaluate different ways to integrate lexical connections into the transformer architecture and present ablation experiments exploring the effect of proposed shortcuts on model behavior.1
1Our code is publicly available to aid the reproduction of the reported results: https://github.com/demelin/ transformer_lexical_shortcuts
Original language | English |
---|---|
Title of host publication | Proceedings of the Fourth Conference on Machine Translation (WMT), Volume 1 |
Subtitle of host publication | Research Papers |
Place of Publication | Florence, Italy |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 102–115 |
Number of pages | 14 |
Volume | 1 |
ISBN (Print) | 978-1-950737-27-7 |
DOIs | |
Publication status | Published - 2 Aug 2019 |
Event | ACL 2019 Fourth Conference on Machine Translation - Florence, Italy Duration: 1 Aug 2019 → 2 Aug 2019 http://www.statmt.org/wmt19/ |
Conference
Conference | ACL 2019 Fourth Conference on Machine Translation |
---|---|
Abbreviated title | WMT19 |
Country/Territory | Italy |
City | Florence |
Period | 1/08/19 → 2/08/19 |
Internet address |
Fingerprint
Dive into the research topics of 'Widening the Representation Bottleneck in Neural Machine Translation with Lexical Shortcuts'. Together they form a unique fingerprint.Projects
- 1 Finished