Is it good data for multilingual instruction tuning or just bad multilingual evaluation for large language models?

Pinzhen Chen, Simon Yu, Zhicheng Guo, Barry Haddow

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract / Description of output

Multilingual large language models are designed, claimed, and expected to cater to speakers of varied languages. We hypothesise that the current practices of fine-tuning and evaluating these models may not perfectly align with this objective owing to a heavy reliance on translation, which cannot cover language-specific knowledge but can introduce translation defects. It remains unknown whether the nature of the instruction data has an impact on the model output; conversely, it is questionable whether translated test sets can capture such nuances. Due to the often coupled practices of using translated data in both stages, such imperfections could have been overlooked. This work investigates these issues using controlled native or translated data during the instruction tuning and evaluation stages. We show that native or generation benchmarks reveal a notable difference between native and translated instruction data especially when model performance is high, whereas other types of test sets cannot. The comparison between round-trip and single-pass translations reflects the importance of knowledge from language-native resources. Finally, we demonstrate that regularization is beneficial to bridging this gap on structured but not generative tasks.
Original languageEnglish
Title of host publicationProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
EditorsYaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
PublisherAssociation for Computational Linguistics
Pages9706-9726
Number of pages21
ISBN (Electronic)9798891761643
Publication statusPublished - 16 Nov 2024
Event2024 Conference on Empirical Methods in Natural Language Processing - Hyatt Regency Miami Hotel, Miami, United States
Duration: 12 Nov 202416 Nov 2024
https://2024.emnlp.org/

Conference

Conference2024 Conference on Empirical Methods in Natural Language Processing
Abbreviated titleEMNLP2024
Country/TerritoryUnited States
CityMiami
Period12/11/2416/11/24
Internet address

Fingerprint

Dive into the research topics of 'Is it good data for multilingual instruction tuning or just bad multilingual evaluation for large language models?'. Together they form a unique fingerprint.

Cite this