TY - GEN
T1 - Knowledge-Graph NLP Assistant for Mitigate Norm Violations
T2 - 2024 Beyond Technology Summit on Informatics International Conference, BTS-I2C 2024
AU - Mulyadi, Hartawan Bahari
AU - Sarno, Riyanarto
AU - Septiyanto, Abdullah Faqih
AU - Sunaryono, Dwi
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The rise of real-time communication through live chat and social media platforms has heightened the need for effective content moderation tools to detect and mitigate norm violations, such as hate speech, abusive language, and offensive slang, thereby ensuring user safety. This study develops and evaluates a knowledge-graph-based natural language processing (NLP) framework aimed at improving the accuracy and efficiency of content moderation in live chats. Several machine learning models were explored, including K-Nearest Neighbors (KNN), Support Vector Machine (SVM), Neural Network (NN), Long Short-Term Memory (LSTM), and Bidirectional LSTM (Bi-LSTM), along with their graphenhanced variants. The models were trained and evaluated on a comprehensive dataset including various norm violations. Results show that the Graph-Bi-LSTM model outperforms the others, achieving an F1-score of 0.84, precision of 0.84, and accuracy of 0.81, indicating its potential to improve harmful content detection. Despite these promising outcomes, the study underscores the need for human oversight to prevent false positives and ensure context-aware decision-making. The findings suggest that combining graph-based NLP models with human moderation tools can enhance the transparency and reliability of content moderation in real-time. Further research should optimize the inference time of models and continuously adapt the system to handle evolving language trends and diverse cultural contexts.
AB - The rise of real-time communication through live chat and social media platforms has heightened the need for effective content moderation tools to detect and mitigate norm violations, such as hate speech, abusive language, and offensive slang, thereby ensuring user safety. This study develops and evaluates a knowledge-graph-based natural language processing (NLP) framework aimed at improving the accuracy and efficiency of content moderation in live chats. Several machine learning models were explored, including K-Nearest Neighbors (KNN), Support Vector Machine (SVM), Neural Network (NN), Long Short-Term Memory (LSTM), and Bidirectional LSTM (Bi-LSTM), along with their graphenhanced variants. The models were trained and evaluated on a comprehensive dataset including various norm violations. Results show that the Graph-Bi-LSTM model outperforms the others, achieving an F1-score of 0.84, precision of 0.84, and accuracy of 0.81, indicating its potential to improve harmful content detection. Despite these promising outcomes, the study underscores the need for human oversight to prevent false positives and ensure context-aware decision-making. The findings suggest that combining graph-based NLP models with human moderation tools can enhance the transparency and reliability of content moderation in real-time. Further research should optimize the inference time of models and continuously adapt the system to handle evolving language trends and diverse cultural contexts.
KW - content moderation
KW - knowledge graph
KW - natural language processing
KW - norm violations
KW - real-time communication
UR - https://www.scopus.com/pages/publications/105003235150
U2 - 10.1109/BTS-I2C63534.2024.10942255
DO - 10.1109/BTS-I2C63534.2024.10942255
M3 - Conference contribution
AN - SCOPUS:105003235150
T3 - 2024 Beyond Technology Summit on Informatics International Conference, BTS-I2C 2024
SP - 445
EP - 450
BT - 2024 Beyond Technology Summit on Informatics International Conference, BTS-I2C 2024
A2 - Wibowo, Ferry Wahyu
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 19 December 2024
ER -