TY - GEN
T1 - ECA control using a single affective User dimension
AU - Charles, Fred
AU - Pecune, Florian
AU - Aranyi, Gabor
AU - Pelachaud, Catherine
AU - Cavazza, Marc
N1 - Publisher Copyright:
© 2015 ACM.
PY - 2015/11/9
Y1 - 2015/11/9
N2 - User interaction with Embodied Conversational Agents (ECA) should involve a significant affective component to achieve realism in communication. This aspect has been studied through different frameworks describing the relationship between user and ECA, for instance alignment, rapport and empathy. We conducted an experiment to explore how an ECA's non-verbal expression can be controlled to respond to a single affective dimension generated by users as input. Our system is based on the mapping of a high-level affective dimension, approach/avoidance, onto a new ECA control mechanism in which Action Units (AU) are activated through a neural network. Since 'approach' has been associated to prefrontal cortex activation, we use a measure of prefrontal cortex left-asymmetry through fNIRS as a single input signal representing the user's attitude towards the ECA. We carried out the experiment with 1 0 subjects, who have been instructed to express a positive mental attitude towards the ECA. In return, the ECA facial expression would reflect the perceived attitude under a neurofeedback paradigm. Our results suggest that users are able to successfully interact with the ECA and perceive its response as consistent and realistic, both in terms of ECA responsiveness and in terms of relevance of facial expressions. From a system perspective, the empirical calibration of the network supports a progressive recruitment of various AUs, which provides a principled description of the ECA response and its intensity. Our findings suggest that complex ECA facial expressions can be successfully aligned with one high-level affective dimension. Furthermore, this use of a single dimension as input could support experiments in the fine-tuning of AU activation or their personalization to user preferred modalities.
AB - User interaction with Embodied Conversational Agents (ECA) should involve a significant affective component to achieve realism in communication. This aspect has been studied through different frameworks describing the relationship between user and ECA, for instance alignment, rapport and empathy. We conducted an experiment to explore how an ECA's non-verbal expression can be controlled to respond to a single affective dimension generated by users as input. Our system is based on the mapping of a high-level affective dimension, approach/avoidance, onto a new ECA control mechanism in which Action Units (AU) are activated through a neural network. Since 'approach' has been associated to prefrontal cortex activation, we use a measure of prefrontal cortex left-asymmetry through fNIRS as a single input signal representing the user's attitude towards the ECA. We carried out the experiment with 1 0 subjects, who have been instructed to express a positive mental attitude towards the ECA. In return, the ECA facial expression would reflect the perceived attitude under a neurofeedback paradigm. Our results suggest that users are able to successfully interact with the ECA and perceive its response as consistent and realistic, both in terms of ECA responsiveness and in terms of relevance of facial expressions. From a system perspective, the empirical calibration of the network supports a progressive recruitment of various AUs, which provides a principled description of the ECA response and its intensity. Our findings suggest that complex ECA facial expressions can be successfully aligned with one high-level affective dimension. Furthermore, this use of a single dimension as input could support experiments in the fine-tuning of AU activation or their personalization to user preferred modalities.
KW - Brain-computer interface
KW - Embodied Conversational Agent
KW - Functional near-infrared spectroscopy (FNIRS)
UR - http://www.scopus.com/inward/record.url?scp=84959264730&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84959264730&partnerID=8YFLogxK
U2 - 10.1145/2818346.2820730
DO - 10.1145/2818346.2820730
M3 - Conference contribution
AN - SCOPUS:84959264730
T3 - ICMI 2015 - Proceedings of the 2015 ACM International Conference on Multimodal Interaction
SP - 183
EP - 190
BT - ICMI 2015 - Proceedings of the 2015 ACM International Conference on Multimodal Interaction
PB - Association for Computing Machinery, Inc
T2 - ACM International Conference on Multimodal Interaction, ICMI 2015
Y2 - 9 November 2015 through 13 November 2015
ER -