Canales, RyanRyanCanalesJain, EaktaEaktaJainJörg, SophieSophieJörg0000-0002-7910-85532025-11-242025-11-242025https://fis.uni-bamberg.de/handle/uniba/110683Eye movement plays an important role in face-to-face communication. In this work, we present a deep learning approach for synthesizing the eye movements of avatars for two-party conversations and evaluate viewer perception of different types of eye motions. We aim to synthesize believable gaze behavior based on head motions and audio features as they would typically be available in virtual reality applications. To this end, we captured the head motion, eye motion, and audio of several two-party conversations and trained an RNN-based model to predict where an avatar looks in a two-person conversational scenario. We evaluated our approach with a user study on the perceived quality of the eye animation and compared our method with other eye animation methods. While our model was not rated highest, our model and our user study lead to a series of insights on model features, viewer perception, and study design that we present.engVirtual realityPerceptionAnimationReal-Time Conversational Gaze Synthesis for Avatarsconferenceobjecturn:nbn:de:bvb:473-irb-110683x