Speech, Voice, Text, and Meaning: A Multidisciplinary Approach to Interview Data through the use of digital tools

Arjan Van Hessen, Silvia Calamai, Henk Van Den Heuvel, Stefania Scagliola, Norah Karrouche, Jeannine Beeken, Louise Corti, Christoph Draxler

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

1 Citation (Scopus)
18 Downloads (Pure)


Interview data is multimodal data: it consists of speech sound, facial expression and gestures, captured in a particular situation, and containing textual information and emotion. This workshop shows how a multidisciplinary approach may exploit the full potential of interview data. The workshop first gives a systematic overview of the research fields working with interview data. It then presents the speech technology currently available to support transcribing and annotating interview data, such as automatic speech recognition, speaker diarization, and emotion detection. Finally, scholars who work with interview data and tools may present their work and discover how to make use of existing technology.

Original languageEnglish
Title of host publicationICMI '20
Subtitle of host publicationProceedings of the 2020 International Conference on Multimodal Interaction
Place of PublicationNew York, NY
PublisherAssociation for Computing Machinery
Number of pages2
ISBN (Electronic)978-1-4503-7581-8
Publication statusPublished - 22 Oct 2020
Event22nd ACM International Conference on Multimodal Interaction, ICMI 2020 - Online, Virtual, Online, Netherlands
Duration: 25 Oct 202029 Oct 2020
Conference number: 22


Conference22nd ACM International Conference on Multimodal Interaction, ICMI 2020
Abbreviated titleICMI
CityVirtual, Online
Internet address


  • Annotation
  • Emotion detection
  • Interview data
  • NLP
  • Speech processing
  • Transcription
  • 22/2 OA procedure

Cite this