Abstract

Music emotions can be seen from the audio and lyrics features. Audio is signal data while lyrics are text data. Combining these two features is needed for detecting music emotions. This research used synchronized dataset of chorus audio and lyrics. Audio features that extracted include dynamics, rhythm, timbre, pitch, and tonality features. While the lyric features that extracted are psycholinguistic, stylistic and statistical features. Audio and lyrics features have preprocessing, data normalization and categorization processes. The normalization process used Min-Max Normalization method and the categorization process uses a Rule Based method. Detection of musical emotions is done by weighting the audio and lyric features of the Naive Bayes probability value. From the weighting of these features, we known that audio feature is a dominant feature then a lyric feature. The weighting ratio is 80% for audio features and 20% for lyric features. The accuracy of system using weighting is 0.774. It increased from the accuracy of system without any weighting.

Original languageEnglish
Title of host publicationProceeding - 6th Information Technology International Seminar, ITIS 2020
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages229-233
Number of pages5
ISBN (Electronic)9781728177267
DOIs
Publication statusPublished - 14 Oct 2020
Event6th Information Technology International Seminar, ITIS 2020 - Virtual, Surabaya, Indonesia
Duration: 14 Oct 202016 Oct 2020

Publication series

NameProceeding - 6th Information Technology International Seminar, ITIS 2020

Conference

Conference6th Information Technology International Seminar, ITIS 2020
Country/TerritoryIndonesia
CityVirtual, Surabaya
Period14/10/2016/10/20

Keywords

  • Audio feature
  • Chorus
  • Lyrics feature
  • Music emotion classification
  • Naive Bayes
  • Weighted

Fingerprint

Dive into the research topics of 'Music emotion detection using weighted of audio and lyric features'. Together they form a unique fingerprint.

Cite this