Abstract
A glut of recent research shows that language models capture linguistic structure. Linzen et al. (2016) found that LSTM-based language models may encode syntactic information sufficient to favor verbs which match the number of their subject nouns. Liu et al. (2018) suggested that the high performance of LSTMs may depend on the linguistic structure of the input data, as performance on several artificial tasks was higher with natural language data than with artificial sequential data.
Such work answers the question of whether a model represents linguistic structure. But how and when are these structures acquired? Rather than treating the training process itself as a black box, we investigate how representations of linguistic structure are learned over time. In particular, we demonstrate that different aspects of linguistic structure are learned at different rates, with part of speech tagging acquired early and global topic information learned continuously.
Such work answers the question of whether a model represents linguistic structure. But how and when are these structures acquired? Rather than treating the training process itself as a black box, we investigate how representations of linguistic structure are learned over time. In particular, we demonstrate that different aspects of linguistic structure are learned at different rates, with part of speech tagging acquired early and global topic information learned continuously.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP |
Place of Publication | Brussels, Belgium |
Publisher | ACL Anthology |
Pages | 328-330 |
Number of pages | 3 |
Publication status | Published - Nov 2018 |
Event | 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP - Brussels, Belgium Duration: 1 Nov 2018 → 1 Nov 2018 https://blackboxnlp.github.io/ |
Conference
Conference | 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP |
---|---|
Country/Territory | Belgium |
City | Brussels |
Period | 1/11/18 → 1/11/18 |
Internet address |