Abstract
Foundational large language models (LLMs) can be instruction-tuned to perform open-domain question answering, facilitating applications like chat assistants. While such efforts are often carried out in a single language, we empirically analyze cost-efficient strategies for multilingual scenarios. Our study employs the Alpaca dataset and machine translations of it to form multilingual data, which is then used to tune LLMs through either low-rank adaptation or full-parameter training. Under a controlled computation budget, comparisons show that multilingual tuning is on par or better than tuning a model for each language. Furthermore, multilingual tuning with downsampled data can be as powerful and more robust. Our findings serve as a guide for expanding language support through instruction tuning.
| Original language | English |
|---|---|
| Title of host publication | Findings of the Association for Computational Linguistics: EACL 2024 |
| Publisher | Association for Computational Linguistics |
| Pages | 1347–1356 |
| Number of pages | 10 |
| Publication status | Published - 17 Mar 2024 |
| Event | The 18th Conference of the European Chapter of the Association for Computational Linguistics - St. Julian’s, Malta Duration: 17 Mar 2024 → 22 Mar 2024 Conference number: 18 https://2024.eacl.org/ |
Conference
| Conference | The 18th Conference of the European Chapter of the Association for Computational Linguistics |
|---|---|
| Abbreviated title | EACL 2024 |
| Country/Territory | Malta |
| City | St. Julian’s |
| Period | 17/03/24 → 22/03/24 |
| Internet address |