Integration of synthetic minority oversampling technique for imbalanced class

Noviyanti Santoso*, Wahyu Wibowo, Hilda Himawati

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

16 Citations (Scopus)

Abstract

In the data mining, a class imbalance is a problematic issue to look for the solutions. It probably because machine learning is constructed by using algorithms with assuming the number of instances in each balanced class, so when using a class imbalance, it is possible that the prediction results are not appropriate. They are solutions offered to solve class imbalance issues, including oversampling, undersampling, and synthetic minority oversampling technique (SMOTE). Both oversampling and undersampling have its disadvantages, so SMOTE is an alternative to overcome it. By integrating SMOTE in the data mining classification method such as Naive Bayes, Support Vector Machine (SVM), and Random Forest (RF) is expected to improve the performance of accuracy. In this research, it was found that the data of SMOTE gave better accuracy than the original data. In addition to the three classification methods used, RF gives the highest average AUC, F-measure, and G-means score.

Original languageEnglish
Pages (from-to)102-108
Number of pages7
JournalIndonesian Journal of Electrical Engineering and Computer Science
Volume13
Issue number1
DOIs
Publication statusPublished - Jan 2019

Keywords

  • Accuracy
  • Data mining
  • Imbalanced class
  • SMOTE

Fingerprint

Dive into the research topics of 'Integration of synthetic minority oversampling technique for imbalanced class'. Together they form a unique fingerprint.

Cite this