JSAI2020

Presentation information

General Session

General Session » J-11 Robot and real worlds

[1Q3-GS-11] Robot and real worlds: Multimodal information

Tue. Jun 9, 2020 1:20 PM - 3:00 PM Room Q (jsai2020online-17)

座長:青島武伸(パナソニック株式会社)

1:20 PM - 1:40 PM

[1Q3-GS-11-01] Category formation of real objects using Multimodal Variational Autoencoder

〇Yuto Yoshida1, Akira Taniguchi1, Kaede Hayashi1, Tadahiro Taniguchi1 (1. Univ. of Ritsumeikan)

Keywords:Deep generative model, Multimodal topic model, Variational autoencoder

We propose a neural network-based unsupervised object categorization method for a robot using multimodal sensor information.
The method is an extension of Multimodal Variational Autoencoder (MVAE). In the proposed method, Dirichlet prior is introduced for giving MVAE a clustering capability in the same way as Multimodal latent Dirichlet allocation (MLDA) that has been used for multimodal object categorization by a robot.
We performed comparative experiments with MLDA using both real objects and synthetic data.
The results show that our proposed model has a reduced computational costs compared to MLDA without deteriorating the classification accuracy.

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password