We propose a new shared task on grading student answers with the goal of enabling well-targeted and flexible feedback in a tutorial dialogue setting. We provide an annotated corpus designed for the purpose, a precise specification for a prediction task and an associated evaluation methodology. The task is feasible but non-trivial, which is demonstrated by creating and comparing three alternative baseline systems. We believe that this corpus will be of interest to the researchers working in tex-tual entailment and will stimulate new developments both in natural language processing in tutorial dialogue systems and textual entailment, contradiction detection and other techniques of interest for a variety of computational linguistics tasks.
|Title of host publication||Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies|
|Place of Publication||Montréal, Canada|
|Publisher||Association for Computational Linguistics|
|Number of pages||11|
|Publication status||Published - 1 Jun 2012|