Unsupervised syntactic chunking with acoustic cues: Computational models for prosodic bootstrapping

John K. Pate, Sharon Goldwater

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Learning to group words into phrases without supervision is a hard task for NLP systems, but infants routinely accomplish it. We hypothesize that infants use acoustic cues to prosody, which NLP systems typically ignore. To evaluate the utility of prosodic information for phrase discovery, we present an HMMbased
unsupervised chunker that learns from only transcribed words and raw acoustic correlates to prosody. Unlike previous work on unsupervised parsing and chunking, we use neither gold standard part-of-speech tags nor punctuation in the input. Evaluated on the Switchboard corpus, our model outperforms
several baselines that exploit either lexical or prosodic information alone, and, despite producing a flat structure, performs competitively with a state-of-the-art unsupervised lexicalized parser, with a substantial advantage in precision. Our results support the hypothesis that acoustic-prosodic cues provide useful evidence about syntactic phrases for languagelearning infants.
Original languageEnglish
Title of host publicationProceedings of the 2nd Workshop on Cognitive Modeling and Computational Linguistics
Publication statusPublished - 2011

Fingerprint Dive into the research topics of 'Unsupervised syntactic chunking with acoustic cues: Computational models for prosodic bootstrapping'. Together they form a unique fingerprint.

Cite this