TY - JOUR
T1 - Sarcasm Detection in Indonesian-English Code-Mixed Text Using Multihead Attention-Based Convolutional and Bi-Directional GRU
AU - Alfan Rosid, Mochamad
AU - Oranova Siahaan, Daniel
AU - Saikhu, Ahmad
N1 - Publisher Copyright:
© 2024 The Authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
PY - 2024
Y1 - 2024
N2 - Detecting sarcasm in text is a very challenging task. Sarcasm often depends on context, tone, and cultural references, which can be difficult for machines to understand. In addition, the increasing occurrence of code-mixing in social media posts poses new challenges in sarcasm detection. Research on sarcasm detection in mixed-code text written in languages other than English is still limited owing to the unavailability of public datasets. To overcome this issue, a dataset was built for sarcasm detection in Indonesian-English mixed-code texts. Furthermore, a hybrid model based on a convolutional neural network (CNN) with multi-head attention and a bi-directional gated recurrent unit (BiGRU), named MHA-CovBi, is proposed for sarcasm detection. In the proposed MHA-CovBi model, a combination of FastText and GloVe word embeddings is utilized to assist the model in understanding and processing texts in different languages. GloVe pretrained word embedding is used for vector representation of English words, while FastText pretrained word embedding is used for vector representation of Indonesian words. Moreover, an auxiliary pragmatic feature illustrating the number of pragmatic markers in tweets was incorporated to enhance detection performance. In addition, this study presents a language detection scheme and transliteration process that can be used to handle languages other than Indonesian and English using Google Translate API. The performance of the proposed model was evaluated through comparative analysis against existing approaches. The proposed model successfully outperformed current state-of-the-art models, achieving an accuracy of 94.60% and F1 score of 94.38%.
AB - Detecting sarcasm in text is a very challenging task. Sarcasm often depends on context, tone, and cultural references, which can be difficult for machines to understand. In addition, the increasing occurrence of code-mixing in social media posts poses new challenges in sarcasm detection. Research on sarcasm detection in mixed-code text written in languages other than English is still limited owing to the unavailability of public datasets. To overcome this issue, a dataset was built for sarcasm detection in Indonesian-English mixed-code texts. Furthermore, a hybrid model based on a convolutional neural network (CNN) with multi-head attention and a bi-directional gated recurrent unit (BiGRU), named MHA-CovBi, is proposed for sarcasm detection. In the proposed MHA-CovBi model, a combination of FastText and GloVe word embeddings is utilized to assist the model in understanding and processing texts in different languages. GloVe pretrained word embedding is used for vector representation of English words, while FastText pretrained word embedding is used for vector representation of Indonesian words. Moreover, an auxiliary pragmatic feature illustrating the number of pragmatic markers in tweets was incorporated to enhance detection performance. In addition, this study presents a language detection scheme and transliteration process that can be used to handle languages other than Indonesian and English using Google Translate API. The performance of the proposed model was evaluated through comparative analysis against existing approaches. The proposed model successfully outperformed current state-of-the-art models, achieving an accuracy of 94.60% and F1 score of 94.38%.
KW - BiGRU
KW - Sarcasm detection
KW - code mixed
KW - convolutional
KW - hybrid word embeddings
KW - multi-head attention
UR - http://www.scopus.com/inward/record.url?scp=85200201222&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2024.3436107
DO - 10.1109/ACCESS.2024.3436107
M3 - Article
AN - SCOPUS:85200201222
SN - 2169-3536
VL - 12
SP - 137063
EP - 137079
JO - IEEE Access
JF - IEEE Access
ER -