We report a mixed-design study on the effect of facial animation method (static, synthesized, or tracked expressions) and its synchronization to speaker audio (in sync or delayed by the method’s inherent latency) on an avatar’s perceived naturalness and plausibility. We created a virtual human for an actor and recorded his spontaneous half-minute responses to conversation prompts. As a simulated immersive interaction, 44 participants unfamiliar with the actor observed and rated performances rendered with the avatar, each with the different facial animation methods. Half of them observed performances in sync and the others with the animation method’s latency. Results show audio synchronization did not influence ratings and static faces were rated less natural and less plausible than animated faces. Notably, synthesized expressions were rated as more natural and more plausible than tracked expressions. Moreover, ratings of verbal behavior naturalness differed in the same way. We discuss implications of these results for avatar-mediated communication.
%0 Conference Paper
%1 kullmann2023facialExpressionComparison
%A Kullmann, Peter
%A Menzel, Timo
%A Botsch, Mario
%A Latoschik, Marc Erich
%B Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems
%C New York, NY, USA
%D 2023
%I Association for Computing Machinery
%K myown xrhub
%P 1--7
%R 10.1145/3544549.3585617
%T An Evaluation of Other-Avatar Facial Animation Methods for Social VR
%U https://downloads.hci.informatik.uni-wuerzburg.de/2023-chi-Kullmann-An-Evaluation-of-Other-Avatar-Facial-Animation-Methods-for-Social-VR.pdf
%X We report a mixed-design study on the effect of facial animation method (static, synthesized, or tracked expressions) and its synchronization to speaker audio (in sync or delayed by the method’s inherent latency) on an avatar’s perceived naturalness and plausibility. We created a virtual human for an actor and recorded his spontaneous half-minute responses to conversation prompts. As a simulated immersive interaction, 44 participants unfamiliar with the actor observed and rated performances rendered with the avatar, each with the different facial animation methods. Half of them observed performances in sync and the others with the animation method’s latency. Results show audio synchronization did not influence ratings and static faces were rated less natural and less plausible than animated faces. Notably, synthesized expressions were rated as more natural and more plausible than tracked expressions. Moreover, ratings of verbal behavior naturalness differed in the same way. We discuss implications of these results for avatar-mediated communication.
%@ 9781450394222
@inproceedings{kullmann2023facialExpressionComparison,
abstract = {We report a mixed-design study on the effect of facial animation method (static, synthesized, or tracked expressions) and its synchronization to speaker audio (in sync or delayed by the method’s inherent latency) on an avatar’s perceived naturalness and plausibility. We created a virtual human for an actor and recorded his spontaneous half-minute responses to conversation prompts. As a simulated immersive interaction, 44 participants unfamiliar with the actor observed and rated performances rendered with the avatar, each with the different facial animation methods. Half of them observed performances in sync and the others with the animation method’s latency. Results show audio synchronization did not influence ratings and static faces were rated less natural and less plausible than animated faces. Notably, synthesized expressions were rated as more natural and more plausible than tracked expressions. Moreover, ratings of verbal behavior naturalness differed in the same way. We discuss implications of these results for avatar-mediated communication.},
added-at = {2023-03-31T10:25:44.000+0200},
address = {New York, NY, USA},
author = {Kullmann, Peter and Menzel, Timo and Botsch, Mario and Latoschik, Marc Erich},
biburl = {https://www.bibsonomy.org/bibtex/2e45b19186a7a4df3a368b344f75f045a/hci-uwb},
booktitle = {Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems},
doi = {10.1145/3544549.3585617},
interhash = {2e86d996b5442ca0d725aa27fe59e248},
intrahash = {e45b19186a7a4df3a368b344f75f045a},
isbn = {9781450394222},
keywords = {myown xrhub},
pages = {1--7},
publisher = {Association for Computing Machinery},
series = {CHI EA '23},
timestamp = {2024-11-21T09:27:11.000+0100},
title = {An Evaluation of Other-Avatar Facial Animation Methods for Social VR},
url = {https://downloads.hci.informatik.uni-wuerzburg.de/2023-chi-Kullmann-An-Evaluation-of-Other-Avatar-Facial-Animation-Methods-for-Social-VR.pdf},
year = 2023
}