JSAI2025

Presentation information

Organized Session

Organized Session » OS-44

[3O4-OS-44b] OS-44

Thu. May 29, 2025 1:40 PM - 3:20 PM Room O (Room 1010)

オーガナイザ:長野 匡隼(電気通信大学),谷口 忠大(京都大学),上田 亮(東京大学大学院),鈴木 麗璽(名古屋大学),中村 友昭(電気通信大学),橋本 敬(北陸先端科学技術大学院大学)

2:20 PM - 2:40 PM

[3O4-OS-44b-03] Multi-Agent Reinforcement Learning based on Variational Bayesian Naming Game

〇Hiroaki Ito1, Masatoshi Nagano1, Keita Fukuoka1, Tomoaki Nakamura1, Akira Taniguchi2, Tadahiro Taniguchi3,2 (1. The University of Electro-Communications, 2. Ritsumeikan University, 3. Kyoto University)

Keywords:Symbol Emergence, Multi-Agent Reinforcement Learning, Variational Inference

To achieve cooperative behavior, humans must infer the purpose and thoughts (internal states) of others. Since direct observation of internal states is difficult, humans estimate these states by communication using symbols, such as language. These symbols emerge uniquely based on the group and purpose, a process known as "emergent communication." Conventional multi-agent reinforcement learning methods based on emergent communication utilize the Metropolis-Hastings naming game, assuming a natural setting where independent agents communicate with each other. However, these methods take high computational costs due to the use of sampling for parameter inference and are limited to two-agent scenarios. In this paper, we propose a novel approach combining the Variational Bayesian Naming Game and the Soft Actor-Critic algorithm to reduce computational costs and enable cooperative learning by multiple agents. In this experiment, using a navigation task, where agents aim to reach a goal without collisions through communication, validate the effectiveness of the proposed method.

Authentication for paper PDF access
A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password