Skip to Main content Skip to Navigation
Conference papers

Multimodal modeling of expressiveness for human-machine interaction

Abstract : Myriad of applications involve the interaction of humans with machines , such as reception agents, home assistants, chatbots or autonomous vehicles' agents. Humans can control the virtual agents by the mean of various modalities including sound, vision, and touch. As the number of these applications increases, a key problem is the requirement of integrating all modalities, to leverage the interaction's quality, as well as the user's experience in the virtual world. In this State-of-the-Art review paper, we discuss about designing engaging virtual agents with expressive gestures and prosody. This paper is part of a work that aims to review the mechanisms that govern multimodal interaction, such as the agent's expressiveness and the adaptation of its behavior, to help remove technological barriers and develop a conversational agent capable of adapting naturally and coherently to its interlocutor.
Complete list of metadatas

Cited literature [40 references]  Display  Hide  Download

https://hal-cnrs.archives-ouvertes.fr/hal-02928055
Contributor : Mireille Fares <>
Submitted on : Wednesday, September 2, 2020 - 11:30:58 AM
Last modification on : Wednesday, September 9, 2020 - 3:16:11 AM

File

WACAI_MireilleFares.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-02928055, version 1

Citation

Mireille Fares, Catherine Pelachaud, Nicolas Obin. Multimodal modeling of expressiveness for human-machine interaction. WACAI 2020, Jun 2021, Île d'Oléron, France. ⟨hal-02928055⟩

Share

Metrics

Record views

21

Files downloads

25