Comparing Automatic and Human Evaluation of Local Explanations for Text Classification

Dong Nguyen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Text classification models are becoming increasingly complex and opaque, however for many applications it is essential that the models are interpretable. Recently, a variety of approaches have been proposed for generating local explanations. While robust evaluations are needed to drive further progress, so far it is unclear which evaluation approaches are suitable. This paper is a first step towards more robust evaluations of local explanations. We evaluate a variety of local explanation approaches using automatic measures based on word deletion. Furthermore, we show that an evaluation using a crowdsourcing experiment correlates moderately with these automatic measures and that a variety of other factors also impact the human judgements.
Original languageEnglish
Title of host publicationThe 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Place of PublicationNew Orleans, Louisiana
PublisherAssociation for Computational Linguistics
Pages1069–1078
Number of pages10
DOIs
Publication statusPublished - 6 Jun 2018
Event16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Hyatt Regency New Orleans Hotel, New Orleans, United States
Duration: 1 Jun 20186 Jun 2018
http://naacl2018.org/

Conference

Conference16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Abbreviated titleNAACL HLT 2018
Country/TerritoryUnited States
CityNew Orleans
Period1/06/186/06/18
Internet address

Fingerprint

Dive into the research topics of 'Comparing Automatic and Human Evaluation of Local Explanations for Text Classification'. Together they form a unique fingerprint.

Cite this