Abstract / Description of output
Speech perception involves storing and integrating sequentially presented items. Recent work in cognitive neuroscience has identified temporal and contextual characteristics in humans' neural encoding of speech that may facilitate this temporal processing. In this study, we simulated similar analyses with representations extracted from a computational model that was trained on unlabelled speech with the learning objective of predicting upcoming acoustics. Our simulations revealed temporal dynamics similar to those in brain signals, implying that these properties can arise without linguistic knowledge. Another property shared between brains and the model is that the encoding patterns of phonemes support some degree of cross-context generalization. However, we found evidence that the effectiveness of these generalizations depends on the specific contexts, which suggests that this analysis alone is insufficient to support the presence of context-invariant encoding.
Original language | English |
---|---|
Title of host publication | Proceedings of the Annual Meeting of the Cognitive Science Society |
Publisher | eScholarship University of California |
DOIs | |
Publication status | Accepted/In press - 5 Apr 2024 |
Event | CogSci 2024: Dynamics of Cognition - Rotterdam, Netherlands Duration: 24 Jul 2024 → 27 Jul 2024 https://cognitivesciencesociety.org/cogsci-2024/ |
Conference
Conference | CogSci 2024 |
---|---|
Abbreviated title | CogSci 2024 |
Country/Territory | Netherlands |
City | Rotterdam |
Period | 24/07/24 → 27/07/24 |
Internet address |
Keywords / Materials (for Non-textual outputs)
- speech processing
- speech representations
- computational model