音声研究
Online ISSN : 2189-5961
Print ISSN : 1342-8675
特集「アジアにおけるコーパス・データ駆動型音声研究」
対面対話における感情の自動推定に向けたマルチモーダル特徴量の検討(<特集>アジアにおけるコーパス・データ駆動型音声研究)
Yoshiko ARIMOTOKazuo OKANOYA
著者情報
ジャーナル フリー

2015 年 19 巻 1 号 p. 53-67

詳細
抄録

To develop an automatic emotion estimation system based on speaker information collected during face-to-face conversation, an extensive exploration of the multimodal features of speakers is required. To satisfy this requirement, a multimodal Japanese dialog corpus with dynamic emotional states was created by recording the vocal and facial expressions and physiological reactions of various speakers. Estimation experiments based on a mixed-effect model and multiple regression analysis were conducted to elucidate the relevant features for speaker-independent and speaker-specific emotion estimation. The results revealed that vocal features were most relevant for speaker-independent emotion estimation, whereas facial features were most relevant for speaker-specific emotion estimation.

著者関連情報
© 2015 日本音声学会
前の記事 次の記事
feedback
Top