Host: The Japanese Society for Artificial Intelligence
Name : The 38th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 38
Location : [in Japanese]
Date : May 28, 2024 - May 31, 2024
The self-attention mechanism (attention mechanism) that Transformers possess internally is being used as an effective method in various fields beyond natural language processing, yet there are many unclear points regarding the interpretation of each attention module. This research proposes a method to analyze the internal behavior of Transformer models by mapping the attention mechanism on a head-by-head basis to linguistic functions, specifically targeting Japanese. Concretely, this involves performing transformations such as swapping tokens that correspond to specific parts of speech, or fixing the vocabulary while only changing the syntactic order, and observing the differences in attention head reactions before and after the transformations. By inputting pairs of sentences that change specific parts of speech or dependency relations and acquiring the differences in attention norm in BERT, it was possible to identify attention heads that are characteristic of specific parts of speech or dependency relations by visualizing these differences