Multimodal Integration for Meeting Group Action Segmentation and Recognition

Marc Al-Hames, Alfred Dielmann, Daniel Gatica-Perez, Stephan Reiter, Steve Renals, Gerhard Rigoll, Dong Zhang

Research output: Chapter in Book/Report/Conference proceedingChapter

Abstract / Description of output

We address the problem of segmentation and recognition of sequences of multimodal human interactions in meetings. These interactions can be seen as a rough structure of a meeting, and can be used either as input for a meeting browser or as a first step towards a higher semantic analysis of the meeting. A common lexicon of multimodal group meeting actions, a shared meeting data set, and a common evaluation procedure enable us to compare the different approaches. We compare three different multimodal feature sets and our modelling infrastructures: a higher semantic feature approach, multi-layer HMMs, a multi-stream DBN, as well as a multi-stream mixed-state DBN for disturbed data.
Original languageEnglish
Title of host publicationMachine Learning for Multimodal Interaction: Second International Workshop, MLMI 2005, Edinburgh, UK, July 11-13, 2005, Revised Selected Papers
Subtitle of host publicationSecond International Workshop, MLMI 2005
EditorsSteve Renals, Samy Bengio
Place of PublicationBerlin, Heidelberg
PublisherSpringer Berlin Heidelberg
Number of pages12
ISBN (Electronic)978-3-540-32550-5
ISBN (Print)978-3-540-32549-9
Publication statusPublished - 2006
EventSecond International Workshop (MLMI 2005) - Edinburgh, United Kingdom
Duration: 11 Jul 200513 Jul 2005

Publication series

NameLecture Notes in Computer Science
PublisherSpringer Berlin Heidelberg
ISSN (Print)0302-9743


WorkshopSecond International Workshop (MLMI 2005)
Country/TerritoryUnited Kingdom


Dive into the research topics of 'Multimodal Integration for Meeting Group Action Segmentation and Recognition'. Together they form a unique fingerprint.

Cite this