JSAI2018

Presentation information

Oral presentation

General Session » [General Session] 2. Machine Learning

[2A2] [General Session] 2. Machine Learning

Wed. Jun 6, 2018 1:20 PM - 3:00 PM Room A (4F Emerald Hall)

座長:石畠 正和(NTTコミュニケーション科学基礎研究所)

1:40 PM - 2:00 PM

[2A2-02] Deep few-shot learning with pseudo example optimization

〇Akisato Kimura1,2, Zoubin Ghahramani2, Koh Takeuchi1, Tomoharu Iwata1, Naonori Ueda1 (1. Nippon Telegraph and Telephone Corporation, 2. University of Cambridge)

Keywords:Neural networks, Few-shot learning, Knowledge distillation

This paper proposes a novel method for training neural networks with a limited amount of training data. Our approach is based on knowledge distillation that transfers knowledge from a deep reference neural network to a shallow target one. The proposed method employs this idea to mimic predictions of non-neural networks reference models that are more robust against overfitting that the target neural network. Different from almost all the previous work for knowledge distillation that requires a large amount of labeled training data, the proposed method requires only a small amount of training data. Instead, we introduce pseudo training data that is optimized as a part of model parameters.