Abstract
Unsupervised commonsense reasoning (UCR) is becoming increasingly popular as the construction of commonsense reasoning datasets is expensive, and they are inevitably limited in their scope. A popular approach to UCR is to fine-tune language models with external knowledge (e.g., knowledge graphs), but this usually requires a large number of training examples. In this paper, we propose to transform the downstream multiple choice question answering task into a simpler binary classification task by ranking all candidate answers according to their reasonableness. To this end, for training the model, we convert the knowledge graph triples into reasonable and unreasonable texts. Extensive experimental results show the effectiveness of our approach on various multiple choice question answering benchmarks. Furthermore, compared with existing UCR approaches using KGs, ours is less data hungry. Our code is available at https://github.com/probe2/BUCA
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics Volume 2: Short Papers |
| Publisher | Association for Computational Linguistics (ACL) |
| Pages | 376-387 |
| Volume | 2 |
| Publication status | Published - 9 Jul 2023 |
| Event | The 61st Annual Meeting of the Association for Computational Linguistics - Westin Harbour Castle, Toronto, Canada Duration: 9 Jul 2023 → 14 Jul 2023 Conference number: 61 https://2023.aclweb.org/ |
Conference
| Conference | The 61st Annual Meeting of the Association for Computational Linguistics |
|---|---|
| Abbreviated title | ACL 2023 |
| Country/Territory | Canada |
| City | Toronto |
| Period | 9/07/23 → 14/07/23 |
| Internet address |