JSAI2023

Presentation information

Poster Session

General Session » Poster session

[4Xin1] Poster session 2

Fri. Jun 9, 2023 9:00 AM - 10:40 AM Room X (Exhibition hall B)

[4Xin1-29] Automatic generation of head motion and facial animation from speech in CG avatar dialogue

〇Yuki Fujioka1, Sei Ueno1, Akinobu Lee1 (1.Nagoya Institute of Technology)

Keywords:motion generation, CG Avatar, multi-modal interaction

In recent years, commnication through avatars has become popular and been expected to apply applications. However, operating the avatar can be burdensome as it requires not only speech but also the use of face, head, and hand motions simultaneously. To reduce the burden on the operator, we propose Speech2motion, a model that automatically generates CG avatar motion from speech. In this work, we focus on the motions in conversation, and the Speech2motion model uses LSTM-based neural networks to predict head motion and facial animation. We recorded 70 minites of motion data along with the speech of one speaker during conversation. We then trained the Speech2motion model using the recorded data. Experimental evaluation shows our proposed model achieves a mean opinion score (MOS) of 3.07 in naturalness of generating the motions.

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password