Abstract
Word2Vec (W2V) and GloVe are popular, fast and efficient word embedding algorithms. Their embeddings are widely used and perform well on a variety of natural language processing tasks. Moreover, W2V has recently been adopted in the field of graph embedding, where it underpins several leading algorithms. However, despite their ubiquity and relatively simple model architecture, a theoretical understanding of what the embedding parameters of W2V and GloVe learn and why that it useful in downstream tasks has been lacking. We show that different interactions between PMI vectors reflect semantic word relationships, such as similarity and paraphrasing, that are encoded in low dimensional word embeddings under a suitable projection, theoretically explaining why embeddings of W2V and GloVe work. As a consequence, we also reveal an interesting mathematical interconnection between the considered semantic relationships themselves.
Original language | English |
---|---|
Title of host publication | Advances in Neural Information Processing Systems (NIPS 2019) |
Publisher | Curran Associates Inc |
Pages | 7465-7475 |
Number of pages | 13 |
Volume | 32 |
Publication status | Published - 14 Dec 2019 |
Event | 33rd Conference on Neural Information Processing Systems - Vancouver Convention Centre, Vancouver, Canada Duration: 8 Dec 2019 → 14 Dec 2019 https://neurips.cc/ |
Conference
Conference | 33rd Conference on Neural Information Processing Systems |
---|---|
Abbreviated title | NeurIPS 2019 |
Country/Territory | Canada |
City | Vancouver |
Period | 8/12/19 → 14/12/19 |
Internet address |