Dimensional speech emotion recognition from speech features and word embeddings by using multitask learning

Bagus Tris Atmaja*, Masato Akagi

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

26 Citations (Scopus)

Abstract

The majority of research in speech emotion recognition (SER) is conducted to recognize emotion categories. Recognizing dimensional emotion attributes is also important, however, and it has several advantages over categorical emotion. For this research, we investigate dimensional SER using both speech features and word embeddings. The concatenation network joins acoustic networks and text networks from bimodal features. We demonstrate that those bimodal features, both are extracted from speech, improve the performance of dimensional SER over unimodal SER either using acoustic features or word embeddings. A significant improvement on the valence dimension is contributed by the addition of word embeddings to SER system, while arousal and dominance dimensions are also improved. We proposed a multitask learning (MTL) approach for the prediction of all emotional attributes. This MTL maximizes the concordance correlation between predicted emotion degrees and true emotion labels simultaneously. The findings suggest that the use of MTL with two parameters is better than other evaluated methods in representing the interrelation of emotional attributes. In unimodal results, speech features attain higher performance on arousal and dominance, while word embeddings are better for predicting valence. The overall evaluation uses the concordance correlation coefficient score of the three emotional attributes. We also discuss some differences between categorical and dimensional emotion results from psychological and engineering perspectives.

Original languageEnglish
Article numbere17
JournalAPSIPA Transactions on Signal and Information Processing
Volume9
DOIs
Publication statusPublished - 19 Feb 2020

Keywords

  • Affective computing
  • Dimensional emotion
  • Feature fusion
  • Multitask learning
  • Speech emotion recognition

Fingerprint

Dive into the research topics of 'Dimensional speech emotion recognition from speech features and word embeddings by using multitask learning'. Together they form a unique fingerprint.

Cite this