Multimodal Emotion Recognition Based on Acoustic and Lexcial Features
CSTR:
Author:
Affiliation:

1.School of Physics and Electronic Engineering, Jiangsu Normal University, Xuzhou 221116, China;2.Kewen College, Jiangsu Normal University, Xuzhou 221116, China;3.School of Linguistic Sciences and Arts, Jiangsu Normal University, Xuzhou 221116, China

Clc Number:

Fund Project:

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    In the speech mode, the OpenSMILE toolbox is used to extract low-level acoustic features from the speech signal. Transformer Encoder is richer to excavate deep features from low level acoustic features and fuses them so as to obtain more useful emotional representation. In the text mode, considering the association between pause and emotion, the speech and text are aligned to obtain the pause information and the pause information is added to the transcript text by pause encoding. The utterance-level lexical features are obtained by the improved DC-BERT model. Then,acoustic features and lexical features are fused and the bi-directional long short-term memory based on attention neural network (BiLSTM-ATT) is used for emotion classification. Finally, this paper compares the effects of three different attention mechanisms integrated into BiLSTM on emotion recognition (local attention, self-attention and multi-headed attention),and local attention is found to be the most effective. In the experiments on IEMOCAP dataset, the method proposed in this paper achieves 78.7% in weighted accuracy for four emotion categories, which is better than the baseline system.

    Reference
    Related
    Cited by
Get Citation

GU Yu, JIN Yun, MA Yong, JIANG Fangjiao, YU Jiajia. Multimodal Emotion Recognition Based on Acoustic and Lexcial Features[J].,2022,37(6):1353-1362.

Copy
Related Videos

Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:January 04,2022
  • Revised:November 07,2022
  • Adopted:
  • Online: November 25,2022
  • Published:
Article QR Code