Adversarial Robustness in Hybrid Quantum-Classical Deep Learning for Botnet DGA Detection

Hatma Suryotrisongko*, Yasuo Musashi, Akio Tsuneda, Kenichi Sugitani

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

This paper aims to contribute to the adversarial defense research gap in the current state-of-the-art of adversarial machine learning (ML) attacks and defense. More specifically, it contributes to the metric measurement of the robustness of artificial intelligence (AI)/ML models against adversarial example attacks, which currently remains an open question in the cybersecurity domain and to an even greater extent for quantum computing-based AI/ML applications. We propose a new adversarial robustness measurement approach which measures the statistical properties (such as the average of the accuracies and t-test results) from the performance results of quantum ML model experiments involving various adversarial perturbation coefficients (attack strength) values. We argue that our proposed approach is suitable for practical use in realizing a quantum-safe world because, in the current noisy intermediate-scale quantum devices (NISQs) era, quantum noise is complex and challenging to model and therefore complicates the measurement task or benchmarking. The second contribution of our study is the novel hardened hybrid quantum-classical deep learning (DL) model for botnet domain generation algorithm (DGA) detection, employing a model hardening adversarial training technique for mitigating new types of unknown DGA adversaries since new cyberattack approaches from the cyber arms race need to be anticipated. Our analysis shows the vulnerability of the hybrid quantum DL model to adversarial example attacks by as much as a 19% average drop in accuracy. We also found the superior performance of our hardened model obtained average accuracy gains as high as 5.9%. Furthermore, we found that the hybrid quantum-classical DL approach gives the benefit of suppressing the negative impact of quantum noises on the classifier’s performance. We demonstrated how to apply our proposed measurement approach in evaluating our novel hybrid quantum DL model and highlighted the adversarial robustness of our model against adversarial example attacks as evidence of the practical implication of our study towards advancing the state of quantum adversarial machine learning research for the quantum-safe world.

Original languageEnglish
Pages (from-to)636-644
Number of pages9
JournalJournal of Information Processing
Volume30
DOIs
Publication statusPublished - 2022

Keywords

  • adversarial ML
  • adversarial defense
  • adversarial training
  • computer security
  • cybersecurity
  • quantum adversarial machine learning
  • quantum computing
  • quantum deep learning

Fingerprint

Dive into the research topics of 'Adversarial Robustness in Hybrid Quantum-Classical Deep Learning for Botnet DGA Detection'. Together they form a unique fingerprint.

Cite this