How Many Languages Make Good Multilingual Instruction Tuning? A Case Study on BLOOM

Shaoxiong Ji, Pinzhen Chen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Instruction tuning a large language model with multiple languages can prepare it for multilingual downstream tasks. Nonetheless, it is yet to be determined whether having a handful of languages is sufficient, or whether the benefits increase with the inclusion of more. By fine-tuning large multilingual models on 1 to 52 languages, we present a case study on BLOOM to understand three pertinent factors affecting performance: the number of languages, language exposure, and similarity between training and test languages. Overall we found that 1) expanding language coverage in multilingual instruction tuning proves to be beneficial; 2) accuracy often significantly boots if the test language appears in the instruction mixture; 3) languages' genetic features correlate with cross-lingual transfer more than merely the number of language but different languages benefit to various degrees.
Original languageEnglish
Title of host publicationProceedings of the 31st International Conference on Computational Linguistics
EditorsOwen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Place of PublicationAbu Dhabi, UAE
PublisherAssociation for Computational Linguistics
Pages2575-2581
Number of pages7
ISBN (Print)9798891761964
Publication statusPublished - 1 Jan 2025
EventThe 31st International Conference on Computational Linguistics - Abu Dhabi, United Arab Emirates
Duration: 19 Jan 202524 Jan 2025
Conference number: 31
http://www.wikicfp.com/cfp/servlet/event.showcfp?copyownerid=90704&eventid=180678

Publication series

NameProceedings – International Conference on Computational Linguistics
PublisherACM
ISSN (Print)2951-2093

Conference

ConferenceThe 31st International Conference on Computational Linguistics
Abbreviated titleCOLING 2025
Country/TerritoryUnited Arab Emirates
CityAbu Dhabi
Period19/01/2524/01/25
Internet address

Fingerprint

Dive into the research topics of 'How Many Languages Make Good Multilingual Instruction Tuning? A Case Study on BLOOM'. Together they form a unique fingerprint.

Cite this