2019年度 人工知能学会全国大会(第33回)

講演情報

国際セッション

国際セッション » [ES] E-2 Machine learning

[2H4-E-2] Machine learning: fusion of models

2019年6月5日(水) 15:20 〜 17:00 H会場 (303+304 小会議室)

座長: 松村 真宏(大阪大学)

16:20 〜 16:40

[2H4-E-2-04] Gradient Descent Optimization by Reinforcement Learning

〇Zhu Yingda1、Hayashi Teruaki1、Ohsawa Yukio1 (1. The University of Tokyo)

キーワード:Deep Neural Network、Gradient Descent、Reinforcement Learning

Gradient descent, which helps to search the global minimum of a complex (high dimension) function, is widely used in the deep neural network to minimize the total loss. The representative methods: stochastic gradient descent (SGD) and ADAM (Kingma & Ba, 2014) are the dominant ones to train neural network today. While some sensitive hyper-parameters like learning rate will affect the descent speed or even the convergence. In previous work, these hyper-parameters are often fixed or set by feedback and experience. I propose using reinforcement learning (RL) to optimize the gradient descent process with neural network feedback as input and hyper-parameter action as output to control these hyper-parameters. The experiment results of using RL based optimizer in both fixed and random start point shows better performance than normal optimizers which are set by default hyper-parameters.