JSAI2018

Presentation information

Poster presentation

General Session » Interactive

[3Pin1] インタラクティブ(1)

Thu. Jun 7, 2018 9:00 AM - 10:40 AM Room P (4F Emerald Lobby)

9:00 AM - 10:40 AM

[3Pin1-35] Action acquisition by Memory Reinforcement Learning useing a prior knowledge

〇Yuna Inamori1, Tsubasa Hirakawa1, Takayoshi Yamashita1, Hironobu Fujiyoshi1, Ryota Kashihawa2, Masaki Inaba2, Naoki Nitanda2 (1. Chubu University, 2. DENSO CORPORATION)

Keywords:Reinforcement Learning, Prior Knowledge, replay buffer

Obtaining a human-level control through reinforcement learning (RL) requires massive training. Furthermore, a deep learning-based RL method such as deep Q network (DQN) is difficult to obtain a stable control. In this paper, we propose a novel deep reinforcement learning method to learn stable controls efficiently. Our approach leverages the technique of experience replay and a replay buffer architecture. We manually create a desirable transition sequence and store the transition in the replay buffer at the beginning of training. This hand-crafted transition sequence enables us to avoid random action selections and optimum local policy. Experimental results on a lane-changing task of autonomous driving show that the proposed method can efficiently acquire a stable control.