人工知能学会論文誌
Online ISSN : 1346-8030
Print ISSN : 1346-0714
ISSN-L : 1346-0714
速報論文
Value行列を手掛かりとした Transformerの分析
吉田 稔松本 和幸北 研二
著者情報
ジャーナル フリー

2023 年 38 巻 2 号 論文ID: 38-2_C-MB7

詳細
抄録

We propose a new method to analyze Transformer language models. In Transformer self-attention modules, attention weights are calculated from the query vectors and key vectors. Then, output vectors are obtained by taking the weighted sum of value vectors. While existing works on analysis of Transformer have focused on attention weights, this work focused on value and output matrices. We obtain joint matrices by multiplying both matrices, and show that the trace of the joint matrices are correlated with word co-occurences.

著者関連情報
© 人工知能学会2023
前の記事 次の記事
feedback
Top