No train but gain: Language arithmetic for training-free language adapters enhancement

Mateusz Klimaszewski, Piotr Andruszkiewicz, Alexandra Birch

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Modular deep learning is the state-of-the-art solution for lifting the curse of multilinguality, preventing the impact of negative interference and enabling cross-lingual performance in Multilingual Pre-trained Language Models. However, a trade-off of this approach is the reduction in positive transfer learning from closely related languages. In response, we introduce a novel method called language arithmetic, which enables training-free post-processing to address this limitation. Extending the task arithmetic framework, we apply learning via addition to the language adapters, transitioning the framework from a multi-task to a multilingual setup. The effectiveness of the proposed solution is demonstrated on three downstream tasks in a MAD-X-based set of cross-lingual schemes, acting as a post-processing procedure. Language arithmetic consistently improves the baselines with significant gains, especially in the most challenging case of zero-shot application. Our code and models are available at https://github.com/mklimasz/language-arithmetic.
Original languageEnglish
Title of host publicationProceedings of the 31st International Conference on Computational Linguistics
EditorsOwen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Place of PublicationStroudsburg, PA, USA
PublisherAssociation for Computational Linguistics (ACL)
Pages11121–11134
Number of pages14
ISBN (Electronic) 9798891761964
Publication statusPublished - 24 Jan 2025
EventThe 31st International Conference on Computational Linguistics - Abu Dhabi, United Arab Emirates
Duration: 19 Jan 202524 Jan 2025
Conference number: 31
http://www.wikicfp.com/cfp/servlet/event.showcfp?copyownerid=90704&eventid=180678

Publication series

NameProceedings of COLING
PublisherAssociation for Computational Linguistics (ACL)
ISSN (Print)1525-2477

Conference

ConferenceThe 31st International Conference on Computational Linguistics
Abbreviated titleCOLING 2025
Country/TerritoryUnited Arab Emirates
CityAbu Dhabi
Period19/01/2524/01/25
Internet address

Fingerprint

Dive into the research topics of 'No train but gain: Language arithmetic for training-free language adapters enhancement'. Together they form a unique fingerprint.

Cite this