Can large language models follow concept annotation guidelines? A case study on scientific and financial domains

Marcio Fonseca, Shay B. Cohen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract / Description of output

Although large language models (LLMs) exhibit remarkable capacity to leverage in-context demonstrations, it is still unclear to what extent they can learn new facts or concept definitions via prompts. To address this question, we examine the capacity of instruction-tuned LLMs to follow in-context concept annotation guidelines for zero-shot sentence labeling tasks. We design guidelines that present different types of factual and counterfactual concept definitions, which are used as prompts for zero-shot sentence classification tasks. Our results show that although concept definitions consistently help in task performance, only the larger models (with 70B parameters or more) have limited ability to work under counterfactual contexts. Importantly, only proprietary models such as GPT-3.5 can recognize nonsensical guidelines, which we hypothesize is due to more sophisticated alignment methods. Finally, we find that FALCON-180B-CHAT is outperformed by LLAMA-2-70B-CHAT is most cases, which indicates that increasing model scale does not guarantee better adherence to guidelines. Altogether, our simple evaluation method reveals significant gaps in concept understanding between the most capable open-source language models and the leading proprietary APIs.
Original languageEnglish
Title of host publicationFindings of the Association for Computational Linguistics ACL 2024
EditorsLun-Wei Ku, Andre Martins, Vivek Srikumar
PublisherAssociation for Computational Linguistics
Pages8027-8042
Number of pages16
ISBN (Electronic)9798891760998
DOIs
Publication statusPublished - 16 Aug 2024
EventThe 62nd Annual Meeting of the Association for Computational Linguistics - Centara Grand and Bangkok Convention Centre at CentralWorld, Bangkok, Thailand
Duration: 11 Aug 202416 Aug 2024
Conference number: 62
https://2024.aclweb.org/

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
PublisherAssociation for Computational Linguistics
ISSN (Print)0736-587X

Conference

ConferenceThe 62nd Annual Meeting of the Association for Computational Linguistics
Abbreviated titleACL 2024
Country/TerritoryThailand
CityBangkok
Period11/08/2416/08/24
Internet address

Fingerprint

Dive into the research topics of 'Can large language models follow concept annotation guidelines? A case study on scientific and financial domains'. Together they form a unique fingerprint.

Cite this