2023年度 人工知能学会全国大会(第37回)

講演情報

国際セッション

国際セッション » IS-1 Knowledge engineering

[2U6-IS-1c] Knowledge engineering

2023年6月7日(水) 17:30 〜 18:50 U会場 (遠隔)

Chair: Akinori Abe (Chiba university)

18:30 〜 18:50

[2U6-IS-1c-04] Adversarial Self-attention Misdirection

Improving vision transformers performance with adversarial pre-training

〇Luiz Henrique Mormille1, Masayasu Atsumi1 (1. Soka Univ.)

[[Online, Working-in-progress]]

キーワード:Vision Transformers, Adversarial Learning, Self-attention

In recent years, the Transformer achieved remarkable results in computer vision related tasks, matching, or even surpassing those of convolutional neural networks. However, to achieve state-of-the-art results, vision transformers rely on large architectures and extensive pre-training on very large datasets. One of the main reasons for this limitation is the fact that vision transformers, whose core is its global self-attention computation, inherently lack inductive biases, with solutions often converging on a local minimum. This work presents a new method to pre-train vision transformers, denoted self-attention misdirection. In this pre-training method, an adversarial U-Net like network pre-processes the input images, altering them with the goal of misdirecting the self-attention computation process in the vision transformer. It uses style representations of image patches to generate inputs that are difficult for self-attention learning, leading the vision transformer to learn representations that generalize better on unseen data.

講演PDFパスワード認証
論文PDFの閲覧にはログインが必要です。参加登録者の方は「参加者用ログイン」画面からログインしてください。あるいは論文PDF閲覧用のパスワードを以下にご入力ください。

パスワード