Decoding Sentiment from Distributed Representations of Sentences

Edoardo Maria Ponti, Ivan Vulić, Anna Korhonen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract / Description of output

Distributed representations of sentences have been developed recently to represent their meaning as real-valued vectors. However, it is not clear how much information such representations retain about the polarity of sentences. To study this question, we decode sentiment from unsupervised sentence representations learned with different architectures (sensitive to the order of words, the order of sentences, or none) in 9 typologically diverse languages. Sentiment results from the (recursive) composition of lexical items and grammatical strategies such as negation and concession. The results are manifold: we show that there is no `one-size-fits-all' representation architecture outperforming the others across the board. Rather, the top-ranking architectures depend on the language at hand. Moreover, we find that in several cases the additive composition model based on skip-gram word vectors may surpass supervised state-of-art architectures such as bi-directional LSTMs. Finally, we provide a possible explanation of the observed variation based on the type of negative constructions in each language.
Original languageEnglish
Title of host publicationProceedings of the 6th Joint Conference on Lexical and Computational Semantics (*SEM 2017)
EditorsNancy Ide, Aurélie Herbelot, Lluís Màrquez
Place of PublicationStroudsburg, PA, USA
PublisherAssociation for Computational Linguistics
Number of pages11
ISBN (Electronic)978-1-945626-53-1
Publication statusPublished - 1 Aug 2017
Event6th Joint Conference on Lexical and Computational Semantics - Vancouver, Canada
Duration: 3 Aug 20174 Aug 2017


Conference6th Joint Conference on Lexical and Computational Semantics
Abbreviated title*SEM 2017
Internet address


Dive into the research topics of 'Decoding Sentiment from Distributed Representations of Sentences'. Together they form a unique fingerprint.

Cite this