Abstract
Recent work on distributional methods for similarity focuses on using the context in which a target word occurs to derive context-sensitive similarity computations. In this paper we present a method for computing similarity which builds vector representations for words in context by modelling senses as latent variables in a large corpus. We apply this to the Lexical Substitution Task and we show that our model significantly outperforms typical distributional methods.
Original language | English |
---|---|
Title of host publication | COLING 2010, 23rd International Conference on Computational Linguistics, Posters Volume, 23-27 August 2010, Beijing, China |
Publisher | Association for Computational Linguistics |
Pages | 250-258 |
Number of pages | 9 |
Publication status | Published - 2010 |