Automatic Multi-Layer Corpus Annotation for Evaluating Question Answering Methods: CBC4Kids

Jochen L. Leidner, Tiphaine Dalmas, Bonnie Webber, Johan Bos, Claire Grover

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Reading comprehension tests are receiving increased attention within the NLP community as a controlled test-bed for developing, evaluating and comparing robust question answering (NLQA) methods. To support this, we have enriched the MITRE CBC4Kids corpus with multiple XML annotation layers recording the output of various tokenizers, lemmatizers, a stemmer, a semantic tagger, POS taggers and syntactic parsers. Using this resource, we have built a baseline NLQA system for wordoverlap based answer retrieval.
Original languageEnglish
Title of host publicationIn Proceedings of the 3rd International Workshop on Linguistically Interpreted Corpora
Pages39-46
Number of pages8
DOIs
Publication statusPublished - 2003

Fingerprint Dive into the research topics of 'Automatic Multi-Layer Corpus Annotation for Evaluating Question Answering Methods: CBC4Kids'. Together they form a unique fingerprint.

Cite this