2:40 PM - 3:00 PM
[2G3-GS-2e-05] Human Motion Forecasting Using GPT-2
Keywords:GPT-2, human motion forecasting, Transformer, Language model, VQ-VAE
Language models such as GPT-2 and BERT improve the performance of language understanding tasks and language generation. These language models have begun to be shown to be applicable not only to language but also to non-linguistic data such as images and audio. By discretizing continuous data using VQ-VAE, it is possible to treat continuous data with language models in the same way as language data. We believe this discretization and learning discrete sequences by the language model can be applied to various types of data. The purpose of this study is to verify the modeling of human motion data using VQ-VAE and GPT-2. In our experiments, we trained VQ-VAE and GPT-2 on the CMU-mocap and 3DPW mocap dataset. We validated the learned models by forecasting the future motion from the motion input of current few frames.
Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.