Abstract
One of many skills required to engage properly in a conversation is to know the appropiate use of the rules of engagement. In order to engage properly in a conversation, a virtual human or robot should, for instance, be able to know when it is being addressed or when the speaker is about to hand over the turn. The paper presents a multimodal approach to end-of-speaker-turn prediction using sequential probabilistic models (Conditional Random Fields) to learn a model from observations of real-life multi-party meetings. Although the results are not as good as expected, we provide insight into which modalities are important when taking a multimodal approach to the problem based on literature and our own results.
Original language | Undefined |
---|---|
Title of host publication | Proceedings of the International Conference on Multimodal Interfaces, ICMI-MLMI 2009 |
Place of Publication | New York |
Publisher | Association for Computing Machinery |
Pages | 91-98 |
Number of pages | 8 |
ISBN (Print) | 978-1-60558-772-1 |
DOIs | |
Publication status | Published - 2009 |
Event | 11th International Conference on Multimodal Interfaces, ICMI 2009 - Boston, United States Duration: 2 Nov 2009 → 4 Nov 2009 Conference number: 11 |
Publication series
Name | |
---|---|
Publisher | ACM |
Conference
Conference | 11th International Conference on Multimodal Interfaces, ICMI 2009 |
---|---|
Abbreviated title | ICMI |
Country/Territory | United States |
City | Boston |
Period | 2/11/09 → 4/11/09 |
Keywords
- METIS-268953
- Machine Learning
- IR-73121
- End-of-Turn Prediction
- EWI-17022
- EC Grant Agreement nr.: FP7/211486
- HMI-IA: Intelligent Agents
- HMI-HF: Human Factors
- Multimodal