人工知能学会論文誌
Online ISSN : 1346-8030
Print ISSN : 1346-0714
ISSN-L : 1346-0714
一般論文
多次元オブジェクト関係の理解のためのテンソル型注意機構
中辻 真藤原 靖宏大塚 淳史野本 済央佐藤 吉秀
著者情報
ジャーナル オープンアクセス HTML

2025 年 40 巻 5 号 p. B-P11_1-12

詳細
抄録

Attention mechanisms have played a crucial role in the success of Transformer models, as seen in platforms likeChatGPT. However, since they compute attentions from relationships between only one or two object types, they failto effectively capture multi-object relationships in real-world scenarios, resulting in low prediction accuracy. In fact,they cannot calculate attention weights among diverse object types, such as the ‘comments,’ ‘replies,’ and ‘subjects’that naturally constitute conversations on platforms like Reddit or X, although their relationships are simultaneouslyobserved in real-world contexts. To overcome this limitation, we introduce the Tensorized Attention Model (TAM),which leverages the Tucker decomposition to calculate attention weights across various object types and seamlesslyintegrates them into the Transformer models. Evaluations show that TAM significantly outperforms existing encodermethods, and its integration into the LoRA adapter for Llama2 enhances fine-tuning accuracy.

著者関連情報
© JSAI (The Japanese Society for Artificial Intelligence)
前の記事 次の記事
feedback
Top