2023年度 人工知能学会全国大会(第37回)

講演情報

国際セッション

国際セッション » IS-2 Machine learning

[2U4-IS-2c] Machine learning

2023年6月7日(水) 13:30 〜 15:10 U会場 (遠隔)

Chair: Tohgoroh Matsui (Chubu university)

14:50 〜 15:10

[2U4-IS-2c-05] Rapid training of Perceiver in a low-cost computing environment

〇Kaiyu Suzuki1, Tomofumi Matsuzawa1 (1. Tokyo University of Science)

[[Online, Regular]]

キーワード:Deep Learning, Image Recognition, Foundation Models, Attention Mechanism

Perceiver is a deep learning model that can be applied to a variety of modalities. It can simultaneously process various forms of input and output, such as images, speech, and natural language using the same architecture. However, Perceiver is computationally more expensive than other models. Therefore, training the model in environments with relatively limited fast parallel computational resources is relatively difficult. In this study, we aimed to reduce the computational cost such that learning can be performed in a short time in environments other than large-scale computing systems. To this end, we first show that a speed-up method proposed for Transformer is also effective for Perceiver. In particular, the gated attention unit proposed for FLASH reduces computational complexity without sacrificing accuracy. The proposed acceleration method can achieve accuracy comparable to that of the original model in a limited computing environment. As an introductory example, we conducted experiments using the ImageNet image recognition task and demonstrated that the proposed method can reduce the training time compared to conventional methods without a significant loss of accuracy. This model can be used to input and output any kind of data quickly in a low-cost computing environment.

講演PDFパスワード認証
論文PDFの閲覧にはログインが必要です。参加登録者の方は「参加者用ログイン」画面からログインしてください。あるいは論文PDF閲覧用のパスワードを以下にご入力ください。

パスワード