Abstract
Continued training is an effective method for domain adaptation in neural machine translation. However, in-domain gains from adaptation come at the expense of general-domain performance. In this work, we interpret the drop in general-domain performance as catastrophic forgetting of general-domain knowledge. To mitigate it, we adapt Elastic Weight Consolidation (EWC)—a machine learning method for learning a new task without forgetting previous tasks. Our method retains the majority of general-domain performance lost in continued training without degrading indomain performance, outperforming the previous state-of-the-art. We also explore the full range of general-domain performance available when some in-domain degradation is acceptable.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers) |
Editors | Jill Burstein, Christy Doran, Thamar Solorio |
Place of Publication | Minneapolis, Minnesota |
Publisher | Association for Computational Linguistics |
Pages | 2062–2068 |
Number of pages | 7 |
Volume | 1 |
DOIs | |
Publication status | Published - 7 Jun 2019 |
Event | 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics - Minneapolis, United States Duration: 2 Jun 2019 → 7 Jun 2019 https://naacl2019.org/ |
Conference
Conference | 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics |
---|---|
Abbreviated title | NAACL-HLT 2019 |
Country/Territory | United States |
City | Minneapolis |
Period | 2/06/19 → 7/06/19 |
Internet address |