Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers

Marc Al-Hamas, Thomas Hain, Jan Cernocky, Sascha Schreiber, Mannes Poel, R.J. Rienks

  • 1 Citations

Abstract

The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?��?. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.
Original languageUndefined
Title of host publicationProceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)
Place of PublicationLondon
PublisherSpringer Verlag
Pages24-35
Number of pages12
ISBN (Print)978-3-540-69267-6
DOIs
StatePublished - 2007

Publication series

NameLecture Notes in Computer Science
PublisherSpringer Verlag
Volume4299
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Fingerprint

Human computer interaction

Keywords

  • IR-63415
  • EC Grant Agreement nr.: FP6/506811
  • METIS-242051
  • EWI-6857

Cite this

Al-Hamas, M., Hain, T., Cernocky, J., Schreiber, S., Poel, M., & Rienks, R. J. (2007). Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers. In Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006) (pp. 24-35). [10.1007/11965152_3] (Lecture Notes in Computer Science; Vol. 4299). London: Springer Verlag. DOI: 10.1007/11965152_3

Al-Hamas, Marc; Hain, Thomas; Cernocky, Jan; Schreiber, Sascha; Poel, Mannes; Rienks, R.J. / Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers.

Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006). London : Springer Verlag, 2007. p. 24-35 10.1007/11965152_3 (Lecture Notes in Computer Science; Vol. 4299).

Research output: Scientific - peer-reviewConference contribution

@inbook{77b5ad81ef69496e91340a7dad0d4ae1,
title = "Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers",
abstract = "The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?��?. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.",
keywords = "IR-63415, EC Grant Agreement nr.: FP6/506811, METIS-242051, EWI-6857",
author = "Marc Al-Hamas and Thomas Hain and Jan Cernocky and Sascha Schreiber and Mannes Poel and R.J. Rienks",
note = "10.1007/11965152_3",
year = "2007",
doi = "10.1007/11965152_3",
isbn = "978-3-540-69267-6",
series = "Lecture Notes in Computer Science",
publisher = "Springer Verlag",
pages = "24--35",
booktitle = "Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)",

}

Al-Hamas, M, Hain, T, Cernocky, J, Schreiber, S, Poel, M & Rienks, RJ 2007, Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers. in Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)., 10.1007/11965152_3, Lecture Notes in Computer Science, vol. 4299, Springer Verlag, London, pp. 24-35. DOI: 10.1007/11965152_3

Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers. / Al-Hamas, Marc; Hain, Thomas; Cernocky, Jan; Schreiber, Sascha; Poel, Mannes; Rienks, R.J.

Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006). London : Springer Verlag, 2007. p. 24-35 10.1007/11965152_3 (Lecture Notes in Computer Science; Vol. 4299).

Research output: Scientific - peer-reviewConference contribution

TY - CHAP

T1 - Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers

AU - Al-Hamas,Marc

AU - Hain,Thomas

AU - Cernocky,Jan

AU - Schreiber,Sascha

AU - Poel,Mannes

AU - Rienks,R.J.

N1 - 10.1007/11965152_3

PY - 2007

Y1 - 2007

N2 - The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?��?. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.

AB - The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?��?. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.

KW - IR-63415

KW - EC Grant Agreement nr.: FP6/506811

KW - METIS-242051

KW - EWI-6857

U2 - 10.1007/11965152_3

DO - 10.1007/11965152_3

M3 - Conference contribution

SN - 978-3-540-69267-6

T3 - Lecture Notes in Computer Science

SP - 24

EP - 35

BT - Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)

PB - Springer Verlag

ER -

Al-Hamas M, Hain T, Cernocky J, Schreiber S, Poel M, Rienks RJ. Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers. In Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006). London: Springer Verlag. 2007. p. 24-35. 10.1007/11965152_3. (Lecture Notes in Computer Science). Available from, DOI: 10.1007/11965152_3