JSAI2024

Presentation information

Poster Session

Poster session » Poster session

[3Xin2] Poster session 1

Thu. May 30, 2024 11:00 AM - 12:40 PM Room X (Event hall 1)

[3Xin2-69] Function Analysis of Attention Heads in Japanese Pretrained BERT Using Linguistically Different Inputs

〇Mizuki Baba1, Yoshinobu Kano1 (1.Shizuoka Univercity)

Keywords:NLP, Transformer, Attention

The self-attention mechanism (attention mechanism) that Transformers possess internally is being used as an effective method in various fields beyond natural language processing, yet there are many unclear points regarding the interpretation of each attention module. This research proposes a method to analyze the internal behavior of Transformer models by mapping the attention mechanism on a head-by-head basis to linguistic functions, specifically targeting Japanese. Concretely, this involves performing transformations such as swapping tokens that correspond to specific parts of speech, or fixing the vocabulary while only changing the syntactic order, and observing the differences in attention head reactions before and after the transformations. By inputting pairs of sentences that change specific parts of speech or dependency relations and acquiring the differences in attention norm in BERT, it was possible to identify attention heads that are characteristic of specific parts of speech or dependency relations by visualizing these differences

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password