The Highs and Lows of Simple Lexical Domain Adaptation Approaches for Neural Machine Translation

Nikolay Bogoychev, Pinzhen Chen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Machine translation systems are vulnerable to domain mismatch, especially in a low-resource scenario. Out-of-domain translations are often of poor quality and prone to hallucinations, due to exposure bias and the decoder acting as a language model. We adopt two approaches to alleviate this problem: lexical shortlisting restricted by IBM statistical alignments, and hypothesis reranking based on similarity. The methods are computationally cheap and show success on low-resource out-of-domain test sets. However, the methods lose advantage when there is sufficient data or too great domain mismatch. This is due to both the IBM model losing its advantage over the implicitly learned neural alignment, and issues with subword segmentation of unseen words.
Original languageEnglish
Title of host publicationProceedings of the Second Workshop on Insights from Negative Results in NLP
EditorsJoão Sedoc, Anna Rogers, Anna Rumshisky, Shabnam Tafreshi
Place of PublicationOnline and Punta Cana, Dominican Republic
PublisherAssociation for Computational Linguistics
Pages74-80
Number of pages7
DOIs
Publication statusPublished - 1 Nov 2021

Fingerprint

Dive into the research topics of 'The Highs and Lows of Simple Lexical Domain Adaptation Approaches for Neural Machine Translation'. Together they form a unique fingerprint.

Cite this