Conference paper

AL-HAMES Marc, HAIN Thomas, ČERNOCKÝ Jan, SCHREIBER Sascha, POEL Mannes, MÜLLER Ronald, MARCEL Sebastien, VAN Leeuwen David, ODOBEZ Jean-Marc, BA Sileye, BOURLARD Herve, CARDINAUX Fabien, GATICA-PEREZ Daniel, JANIN Adam, MOTLÍČEK Petr, REITER Stephan, RENALS Steve, VAN Rest Jeroen, RIENKS Rutger, RIGOLL Gerhard, SMITH Kevin, THEAN Andrew and ZEMČÍK Pavel. Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers. In: Proc. 3nd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006). Washington D.C., 2006, p. 12.
Publication language:english
Original title:Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers
Title (cs):Audiovizuální zpracování meetingů - sedm otázek a odpovědí projektu AMI
Pages:12
Proceedings:Proc. 3nd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)
Conference:3nd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms
Place:Washington D.C., US
Year:2006
URL:http://www.fit.vutbr.cz/~cernocky/publi/2006/wp4_mlmi_final.pdf [PDF]
Keywords
speech processing, video processing, multi-modal interaction
Annotation
The paper is on Audio-Visual Processing in Meetings: it asks Seven Questions and presents Current AMI Answers
Abstract
The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms - and the required component technologies R and D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like "Who is acting during the meeting?". We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions
BibTeX:
@INPROCEEDINGS{
   author = {Marc Al-Hames and Thomas Hain and Jan {\v{C}}ernock{\'{y}}
	and Sascha Schreiber and Mannes Poel and Ronald M{\"{u}}ller
	and Sebastien Marcel and David Leeuwen van and Jean-Marc
	Odobez and Sileye Ba and Herve Bourlard and Fabien Cardinaux
	and Daniel Gatica-Perez and Adam Janin and Petr
	Motl{\'{i}}{\v{c}}ek and Stephan Reiter and Steve Renals and
	Jeroen Rest van and Rutger Rienks and Gerhard Rigoll and
	Kevin Smith and Andrew Thean and Pavel Zem{\v{c}}{\'{i}}k},
   title = {Audio-Visual Processing in Meetings: Seven Questions and
	Current AMI Answers},
   pages = {12},
   booktitle = {Proc. 3nd Joint Workshop on Multimodal Interaction and
	Related Machine Learning Algorithms (MLMI 2006)},
   year = {2006},
   location = {Washington D.C., US},
   language = {english},
   url = {http://www.fit.vutbr.cz/research/view_pub.php?id=8237}
}

Your IPv4 address: 54.156.92.243
Switch to IPv6 connection

DNSSEC [dnssec]