The unreasonable effectiveness of large language models for low-resource clause-level morphology: In-context generalization or prior exposure?

Coleman Haley

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

This paper describes the submission of Team “Giving it a Shot” to the AmericasNLP 2024 Shared Task on Creation of Educational Materials for Indigenous Languages. We use a simple few-shot prompting approach with several state of the art large language models, achieving competitive performance on the shared task, with our best system placing third overall. We perform a preliminary analysis to determine to what degree the performance of our model is due to prior exposure to the task languages, finding that generally our performance is better explained as being derived from in-context learning capabilities.
Original languageEnglish
Title of host publicationProceedings of the 4th Workshop on Natural Language Processing for Indigenous Languages of the Americas (AmericasNLP 2024)
EditorsManuel Mager, Abteen Ebrahimi, Shruti Rijhwani, Arturo Oncevay, Luis Chiruzzo, Robert Pugh, Katharina von der Wense
PublisherAssociation for Computational Linguistics (ACL)
Pages174-178
Number of pages5
ISBN (Electronic)9798891761087
DOIs
Publication statusPublished - 21 Jun 2024
EventThe 4th Workshop on NLP for Indigenous Languages of the Americas - Mexico City, Mexico
Duration: 21 Jun 202421 Jun 2024

Workshop

WorkshopThe 4th Workshop on NLP for Indigenous Languages of the Americas
Abbreviated titleAmericasNLP 2024
Country/TerritoryMexico
CityMexico City
Period21/06/2421/06/24

Fingerprint

Dive into the research topics of 'The unreasonable effectiveness of large language models for low-resource clause-level morphology: In-context generalization or prior exposure?'. Together they form a unique fingerprint.

Cite this