Abstract
Evaluation is crucial in assessing the effectiveness of new information retrieval and human computer interaction techniques and systems. Relevance judgements are often performed by humans, which makes obtaining them expensive and time consuming. Consequently, relevance judgements are usually performed only on a subset of a given collection of data or experimental results with a focus on the top ranked documents. However, when assessing the performance of exploratory search systems, the diversity or subjective relevance of documents that the user was presented with over a search session are often of more importance than the relative ranking of top documents. In order to perform these types of assessment, all the documents in a given collection need to be judged for relevance. In this paper, we propose an approach based on topic modeling that can greatly accelerate document relevance judgment of an entire document collection with an expert assessor needing to mark only a small subset of documents from a given collection. Experimental results show a substantial overlap between relevance judgments compared to a human assessor.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2017 Conference on Conference Human Information Interaction and Retrieval |
Place of Publication | New York, NY, USA |
Publisher | ACM |
Pages | 313-316 |
Number of pages | 4 |
ISBN (Print) | 978-1-4503-4677-1 |
DOIs | |
Publication status | Published - 7 Mar 2017 |
Event | 2017 Conference on Conference Human Information Interaction and Retrieval - Oslo, Norway Duration: 7 Mar 2017 → 11 Mar 2017 http://sigir.org/chiir2017/ |
Publication series
Name | CHIIR '17 |
---|---|
Publisher | ACM |
Conference
Conference | 2017 Conference on Conference Human Information Interaction and Retrieval |
---|---|
Abbreviated title | CHIIR 2017 |
Country/Territory | Norway |
City | Oslo |
Period | 7/03/17 → 11/03/17 |
Internet address |