Measuring Distributional Similarity in Context

Georgiana Dinu, Mirella Lapata

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

The computation of meaning similarity as operationalized by vector-based models has found widespread use in many tasks ranging from the acquisition of synonyms and para-phrases to word sense disambiguation and textual entailment. Vector-based models are typically directed at representing words in isolation and thus best suited for measuring similarity out of context. In his paper we propose a probabilistic framework for measuring similarity in context. Central to our approach is the intuition that word meaning is represented as a probability distribution over a set of latent senses and is modulated by context. Experimental results on lexical substitution and word similarity show that our algorithm outperforms previously proposed models
Original languageEnglish
Title of host publicationEMNLP '10 Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing
PublisherAssociation for Computational Linguistics
Pages1162-1172
Number of pages11
Publication statusPublished - 2010

Fingerprint

Dive into the research topics of 'Measuring Distributional Similarity in Context'. Together they form a unique fingerprint.

Cite this