Huda Ali Alatwi
Realism versus Performance for Adversarial Examples Against DL-based NIDS
Alatwi, Huda Ali; Morisset, Charles
Abstract
The application of deep learning-based (DL) network intrusion detection systems (NIDS) enables effective automated detection of cyberattacks. Such models can extract valuable features from high-dimensional and heterogeneous network traffic with minimal feature engineering and provide high accuracy detection rates. However, it has been shown that DL can be vulnerable to adversarial examples (AEs), which mislead classification decisions at inference time, and several works have shown that AEs are indeed a threat against DL-based NIDS. In this work, we argue that these threats are not necessarily realistic. Indeed, some general techniques used to generate AE manipulate features in a way that would be inconsistent with actual network traffic. In this paper, we first implement the main AE attacks selected from the literature (FGSM, BIM, PGD, NewtonFool, CW, DeepFool, EN, Boundary, HSJ, ZOO) for two different datasets (WSN-DS and BoT-IoT) and we compare their relative performance. We then analyze the perturbation generated by these attacks and use the metrics to establish a notion of "attack unrealism". We conclude that, for these datasets, some of these attacks are performant but not realistic.
Citation
Alatwi, H. A., & Morisset, C. (2023, March). Realism versus Performance for Adversarial Examples Against DL-based NIDS. Presented at SAC '23: 38th ACM/SIGAPP Symposium on Applied Computing, Tallinn Estonia
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | SAC '23: 38th ACM/SIGAPP Symposium on Applied Computing |
Start Date | Mar 27, 2023 |
End Date | Mar 31, 2023 |
Acceptance Date | Mar 7, 2023 |
Online Publication Date | Jun 7, 2023 |
Publication Date | Mar 27, 2023 |
Deposit Date | Jan 20, 2025 |
Publisher | Association for Computing Machinery (ACM) |
Peer Reviewed | Peer Reviewed |
Pages | 1549-1557 |
DOI | https://doi.org/10.1145/3555776.3577671 |
Public URL | https://durham-repository.worktribe.com/output/3342477 |