Audio information access from meeting rooms

Steve Renals, Dan Ellis

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

We investigate approaches to accessing information from the streams of audio data that result from multi-channel recordings of meetings. The methods investigated use word-level transcriptions, and information derived from models of speaker activity and speaker turn patterns. Our experiments include spoken document retrieval for meetings, automatic structuring of meetings based on self-similarity matrices of speaker turn patterns and a simple model of speaker activity. Meeting recordings are rich in both lexical and non-lexical information; our results illustrate some novel kinds of analysis made possible by a transcribed corpus of natural meetings.
Original languageEnglish
Title of host publication2003 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings of the
Subtitle of host publicationICASSP'03
PublisherInstitute of Electrical and Electronics Engineers
Pages744-747
Volume4
ISBN (Print)0-7803-7663-3
DOIs
Publication statusPublished - 2003
Event2003 IEEE International Conference on Acoustics, Speech, and Signal Processing - Hong Kong Exhibition and Convention Centre, Hong Kong, Hong Kong
Duration: 6 Apr 200310 Apr 2003

Conference

Conference2003 IEEE International Conference on Acoustics, Speech, and Signal Processing
Country/TerritoryHong Kong
CityHong Kong
Period6/04/0310/04/03

Fingerprint

Dive into the research topics of 'Audio information access from meeting rooms'. Together they form a unique fingerprint.

Cite this