Multimodal modeling of expressiveness for human-machine interaction

Fiche du document

Date

3 juin 2020

Type de document
Périmètre
Langue
Identifiants
Collection

Archives ouvertes

Licence

info:eu-repo/semantics/OpenAccess




Citer ce document

Mireille Fares et al., « Multimodal modeling of expressiveness for human-machine interaction », HAL-SHS : sociologie, ID : 10670/1.421oc3


Métriques


Partage / Export

Résumé En

Myriad of applications involve the interaction of humans with machines, such as reception agents, home assistants, chatbots or autonomous vehicles’ agents. Humans can control the virtual agents by the mean of various modalities including sound, vision, and touch. As the number of these applications increases, a key problem is the requirement of integrating all modalities, to leverage the interaction’s quality, as well as the user’s experience in the virtual world. In this State-of-the-Art review paper, we discuss about designing engaging virtual agents with expressive gestures and prosody. This paper is part of a work that aims to review the mechanisms that govern multimodal interaction, such as the agent’s expressiveness and the adaptation of its behavior, to help remove technological barriers and develop a conversational agent capable of adapting naturally and coherently to its interlocutor.

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Exporter en