Edinburgh Research Explorer

Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers

Research output: Chapter in Book/Report/Conference proceedingChapter

  • Marc Al-Hames
  • Thomas Hain
  • Jan Cernocky
  • Sascha Schreiber
  • Mannes Poel
  • Ronald Müller
  • Sebastien Marcel
  • David van Leeuwen
  • Jean-Marc Odobez
  • Sileye Ba
  • Herve Bourlard
  • Fabien Cardinaux
  • Daniel Gatica-Perez
  • Adam Janin
  • Petr Motlicek
  • Stephan Reiter
  • Jeroen van Rest
  • Rutger Rienks
  • Gerhard Rigoll
  • Kevin Smith
  • Andrew Thean
  • Pavel Zemcik

Related Edinburgh Organisations

Original languageEnglish
Title of host publicationMachine Learning for Multimodal Interaction
Subtitle of host publicationThird International Workshop, MLMI 2006, Bethesda, MD, USA, May 1-4, 2006, Revised Selected Papers
EditorsSteve Renals, Samy Bengio, Jonathan G. Fiscus
Place of PublicationBerlin, Heidelberg
PublisherSpringer Berlin Heidelberg
Pages24-35
Number of pages12
ISBN (Electronic)978-3-540-69268-3
ISBN (Print)978-3-540-69267-6
DOIs
Publication statusPublished - 2006
EventThird International Workshop MLMI 2006 - Bethesda, MD, United States
Duration: 1 May 20064 May 2006

Publication series

NameLecture Notes in Computer Science
PublisherSpringer Berlin Heidelberg
Volume4299
ISSN (Print)0302-9743

Workshop

WorkshopThird International Workshop MLMI 2006
CountryUnited States
CityBethesda, MD
Period1/05/064/05/06

Abstract

The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?”. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.

Event

Third International Workshop MLMI 2006

1/05/064/05/06

Bethesda, MD, United States

Event: Workshop

ID: 27397709