0% found this document useful (0 votes)
4 views

Adversarial Machine Learning for Network Intrusion Detection a Monsieur Mohamed

This document presents a comparative study on the use of adversarial machine learning for network intrusion detection systems (NIDS), focusing on the robustness of shallow classifiers against various adversarial attacks. The study evaluates seven shallow machine learning models and a deep learning network, highlighting the impact of adversarial attacks and the effectiveness of a Gaussian data augmentation defense technique. The findings emphasize the importance of assessing the vulnerability of NIDS to adversarial attacks to enhance their robustness and security in real-world applications.

Uploaded by

khaoulaajroudi6
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views

Adversarial Machine Learning for Network Intrusion Detection a Monsieur Mohamed

This document presents a comparative study on the use of adversarial machine learning for network intrusion detection systems (NIDS), focusing on the robustness of shallow classifiers against various adversarial attacks. The study evaluates seven shallow machine learning models and a deep learning network, highlighting the impact of adversarial attacks and the effectiveness of a Gaussian data augmentation defense technique. The findings emphasize the importance of assessing the vulnerability of NIDS to adversarial attacks to enhance their robustness and security in real-world applications.

Uploaded by

khaoulaajroudi6
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 30

Adversarial machine learning for network intrusion

detection: a comparative study


Houda Jmila, Mohamed Ibn Khedher

To cite this version:


Houda Jmila, Mohamed Ibn Khedher. Adversarial machine learning for network intrusion
detection: a comparative study. Computer Networks, 2022, 214, pp.109073:1-109073:14.
�10.1016/j.comnet.2022.109073�. �hal-03726213�

HAL Id: hal-03726213


https://hal.science/hal-03726213v1
Submitted on 11 Jan 2024

HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est


archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents
entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non,
lished or not. The documents may come from émanant des établissements d’enseignement et de
teaching and research institutions in France or recherche français ou étrangers, des laboratoires
abroad, or from public or private research centers. publics ou privés.

Distributed under a Creative Commons Attribution - NonCommercial 4.0 International License


Adversarial Machine Learning for Network intrusion Detection: A
Comparative Study

Houda Jmilaa , Mohamed Ibn Khedherb


a Samovar, CNRS, Télécom SudParis, Institut Polytechnique de Paris
9 rue Charles Fourier, 91011 Evry Cedex, France
b IRT-SystemX, 8 Avenue de la Vauve, 91120 Palaiseau, France

Abstract
Intrusion detection is a key topic in cybersecurity. It aims to protect computer systems and networks
from intruders and malicious attacks. Traditional intrusion detection systems (IDS) follow a signature-
based approach, but in the last two decades, various machine learning (ML) techniques have been strongly
proposed and proven to be effective. However, ML faces several challenges, one of the most interesting
being the emergence of adversarial attacks to fool the classifiers. Addressing this vulnerability is critical to
prevent cybercriminals from exploiting ML flaws to bypass IDS and damage data and systems.
Some research papers have studied the vulnerability of ML based IDS to adversarial attacks, however
most of them focused on deep learning based classifiers. Unlike them, this paper pays more attention to
shallow classifiers that are still widely used in ML-based IDS due to their maturity and simplicity of im-
plementation. In more detail, we evaluate the robustness of 7 shallow ML-based NIDS including Adaboost,
Bagging, Gradient boosting (GB), Logistic Regression (LR), Decision Tree (DT), Random Forest (RF), Sup-
port Vector Classifier (SVC) and also a Deep Learning Network, against several adversarial attacks widely
used in the state of the art (SOA). In addition, we apply a Gaussian data augmentation defence technique and
measure its contribution to improving classifier robustness. We conduct extensive experiments in different
scenarios using the NSL-KDD benchmark dataset [5] and the UNSW-NB 15 dataset [50]. The results show
that attacks do not have the same impact on all classifiers and that the robustness of a classifier depends on
the attack and that a trade-off between performance and robustness must be considered depending on the
network intrusion detection scenario.
Keywords: IDS, Anomaly detection, Adversarial attack, Defence technique, NSL-KDD.

Nomenclature

AAC Anomaly prediction ACcuracy

AdvML Adversarial Machine Learning


ANN Artificial Neural Network
BIM Basic Iterative Method

C&W Carlini and Wagner


CNN Convolutional Neural Networks

Preprint submitted to Elsevier January 11, 2024


DDoS Distributed Denial-of-Service
DNN Deep Neural Network
DT Decision Tree
FFNN Feedforward Neural Network

FGSM Fast Gradient Sign Method


FNR False Negative Rate
GAN Generative Adversarial Networks

GAN Generative adversarial network


GB Gradient Boosting
GMM Gaussian Mixture Model
HIDS Home based IDS

HSJ Hop Skip Jump


IDS Intrusion Detection System
IoT Internet Of Things

JSMA Jacobian based Saliency Map Attack


KNN K-Nearest Neighbors
ML Machine Learning
NIDS Network-based IDS

PGD Projected Gradient Descent


RF Reinforcement Learning
RNN Recurrent Neural Network

SOA State Of the Art


SVC Support Vector classifier
SVM Support Vector Machine
TAC Total prediction ACcuracy

UAP Universal Adversarial Perturbations


ZOO Zeroth-order optimization

2
1. Introduction

Protecting computer systems and networks from cyberattacks has been a growing concern in recent
years. Although most systems are built with improved security features, a large number of vulnerabilities
still exist. These include unwanted access to systems and information, destruction or alteration of data, etc.
Intrusion detection systems play a critical role in the network defence process and allow network operators
to accurately identify security attacks. There are mainly two categories of IDS: Network-based IDS and
Host-based IDS, described below:

• Network-based IDSs (NIDS) monitor and analyze network traffic at different layers to detect intruders.
• Host-based IDS (HIDS), monitor the computer infrastructure to detect internal changes by exploiting
host indicators such as sensor log files, disk resources, user account information processes, etc.

This paper focuses on the Network-based IDSs. The continuous increase in the number and types of
contemporary network threats [49] motivates this interest.
Both NIDS and HIDS approaches can be classified into the following categories:
• Misuse-based approaches (also called signature-based) exploit indicators (or signatures) previously
extracted from known attacks. Signatures are manually generated for each new attack. Therefore,
maintaining an up-to-date list of signatures is costly due to the increasing number and diversity of
attacks.
• Anomaly-based approaches model normal network behavior, as opposed to malicious behavior. Al-
though these approaches are capable of detecting new attacks, they suffer from a high false alarm rate
because new normal behavior can be detected as malicious.
Anomaly detection is often considered by the community to be more promising than signature-based detec-
tion, as it is able to detect unknown attacks. Therefore, this paper focuses on anomaly-based NIDS.
In recent years, ML approaches have been widely used for anomaly detection. Existing approaches can
be classified into shallow (or classic) models [14] and deep learning models [23]. Deep learning involves
several levels of representation and several layers of non-linear processing units. On the contrary, all non-
deep learning approaches can be qualified as shallow learning, this includes the majority of conventional
machine learning models proposed prior to 2006 and neural networks with only one hidden layer of nodes
[74]. The most popular shallow approaches include Random Forest (RF), Decision Tree(DT), Support Vector
Machine (SVM), k-Nearest Neighbors (KNN), Hidden Markov Models (HMM) and Ensemble Learning.
Both shallow and deep learning models have been used with promising results.
While most research focuses on designing new ML-based IDSs, this paper highlights the vulnerabilities
of ML systems to adversarial attacks. Adversarial attacks allow a small and carefully designed change in
the input of the ML classifier to completely alter the output of the system. Adversarial Machine Learning
(AdvML) is the research area that studies these vulnerabilities. It has been widely explored in recent years,
particularly in the field of computer vision [73]. The study of AdvML in cybersecurity also deserves a great
deal of attention given the sensitivity of this field and the need to preserve the confidentiality, integrity, and
availability of data and systems. It is essential to evaluate the robustness of ML-based intrusion detection
systems before deploying them in the network. This prevents cyber criminals from exploiting ML vulner-
abilities to bypass IDS and damage data and systems. The robustness of an ML classifier is defined as its
ability to maintain its accuracy against adverse samples. An adverse sample is an input instance with a small
disturbance that is erroneously predicted. Depending on the results of the robustness assessment, appropriate
defence techniques can be applied to improve the robustness of NIDS.

3
Due to the widespread adoption of deep learning approaches for NIDS, most research work evaluate the
robustness of deep learning based NIDS [62]. However, shallow ML models are still widely used in NIDS
due to their simplicity and implementation maturity [66]. It is therefore interesting to study their robustness
in an adversarial environment. This paper focuses on the evaluation of shallow ML based NIDS against
several adversarial attacks widely used in the state of the art.
In this paper, we evaluate the robustness of 7 shallow classifiers including Adaboost, Bagging, Gradi-
ent boosting, Logistic regression, Decision Tree, random forest, Support Vector Classifier and also a Deep
Learning Network, against a wide range of attacks (an attack is defined as a method of generating adver-
sarial examples). In particular, we consider white-box and gray/black-box attacks. In white-box attacks the
attacker has full access to all information about the ML-based NIDS, whereas in gray/black-box attacks, the
attacker has little or no knowledge of ML-based NIDS. Gray/black-box attacks are interesting because they
represent the most realistic scenario for adversary’s attacks. Examining white-box attacks is useful for IDS
manufacturers who has full access to their system and wish to evaluate its performance against adversarial
attacks.
This document provides the following main contributions:

• A clear and structured survey of most commonly used adversarial attacks and defence techniques, in
addition to an exhaustive review of current work on Adversarial ML NIDS.
• An in-depth study of the impact of adversarial attacks on ML based NIDS. Several types of attacks (9
white-box and gray/black-box attacks) are explored with a particular attention to shallow classifiers.
Indeed, unlike the overwhelming majority of works that study the behavior of NIDS in an adversarial
environment and focus on deep learning approaches, this paper focuses on shallow algorithms, which
are still widely used in ML-based NIDS thanks to their simplicity of implementation and maturity.
The evaluation of their performance in an adversarial environment is therefore also worth exploring.
• An evaluation of the contribution of a Gaussian data augmentation defence technique to improving
the robustness of the classifiers.
• Valuable results and conclusions that can help security researchers improve the robustness of their
NIDS. These results are deduced based on extensive experiments conducted under different scenarios.
• The steps in the study conducted represent a framework detailing the steps to be taken to assess the
sensitivity of NIDS to adversary attacks and improve their robustness.

The paper is structured as follows. Section 2 describes the challenges in the field of network intrusion
detection. Section 3, provides a state of the art of the most commonly used adversarial attacks and defence
techniques, as well as an exhaustive study of AdvML approaches in the field of NIDS. Section 4 describes
our evaluation study, including the evaluation parameters and protocol. Section 5 details the experimental
results. Section 6 provides a discussion and section 7 concludes the paper.

2. The Challenging task of Network Intrusion Detection

Network intrusion detection is a complex task for many reasons. Challenges can be related to the nature
of the network traffic data, or to the inherent NIDS decision model, as described below.

4
Figure 1: Adversarial attack generation in NIDS

An adverse sample is generated by adding a small perturbation to the original sample. Thus, malicious perturbed traffic
can be misclassified as benign and thus bypass the intrusion detection system. This can have serious consequences for the
system.

2.1. Challenges related to the nature of the network traffic data:


• Challenges related to data exploitation: in network anomaly detection, the captured packets partially
represent the entire network traffic because the observation points are often widely distributed. Thus,
the captured data is often sparse, huge and contains redundant or uninformative data, which makes it
difficult to exploit.
• Unbalanced data-sets: In most IDS data-sets, the amount of normal data dominates the data. This is
due to the low frequency of attacks compared to normal behavior. This makes the available data-sets
unbalanced. Therefore, classical machine learning algorithms need to be adapted to the context of
unbalanced data-sets.
• Variety of attacks: the attack landscape is frequently changing as attackers are constantly developing
new methods. Attack detection and analysis tools must be updated and evolve continuously.

2.2. Challenges related to the decision model


• Scalability: this is a common challenge for statistical learning algorithms. It is defined by the ability
of the algorithm to function normally even with high dimensional data.
• Real-time IDS: The goal of an IDS is to detect attacks and stop them before they damage the system.
Therefore, the design of a real-time IDS is very important. A real-time IDS must also be efficient and
flexible to run on most commercial computers.
• High false positive rate: This is the reporting of a high number of false alarms that correspond to
legitimate activity that has been misclassified by the IDS. Recognizing true alarms from the huge
volume of alarms is a complicated and time consuming task. Therefore, reducing false alarms is a
serious problem to ensure the effectiveness and use of IDS.
• Vulnerability to Adversarial Attacks: This aspect has been described above and is the focus of this pa-
per. The objective is to examine and reduce the vulnerability of various NIDS classifiers to adversarial
attacks.

3. Adversarial attacks and defence techniques in NIDS: Background and Review

3.1. Preliminaries
Generating an adversarial attack involves adding a small perturbation to the input sample so that the
output label is misclassified. This is illustrated in Figure 1 in the context of NIDS. Formally, let x be the

5
original input data sample, f be the classifier, and y = f (x) be the label associated with x. A data sample x′
is considered an adverse sample of x when x′ is close to x under a specific distance metric while f (x′ ) , y.
Adversarial attacks in network security can be classified along two dimensions: the attacker’s knowledge
and the attacker’s goal :
1. The attacker’s knowledge : describes the extent of the adversary’s knowledge about the NIDS system.
We can characterize three levels of attack danger [29]:

• White-box attacks: the attacker is in the most favorable position where he has full access to
all information about the ML-based NIDS. This includes training data and the learning model
architecture, decision and parameters (gradient, loss function, etc.). Fortunately, this is generally
not feasible in the majority of real adversarial attacks.
• Black-box attacks: This is the opposite case where the attacker completely ignores the ML-based
NIDS system and its inputs/outputs. It can be argued that a truly black-box attack is impossible
and rarely succeeds.
• Gray-box attacks: this scenario assumes a more realistic approach, where the attacker has some
level of knowledge of the ML-based NIDS, and may have limited access to the training data
. The adversary does not have the exact information but has enough information to be able to
attack the ML system and cause it to fail.

Note that in the literature, by abuse of language, the term "black-box attacks" is also used for "gray-
box attacks" (for example, the ZOO attack is called back-box attack in [18]). In this article, we use
the terms "gray/black-box" to refer to the gray-box attacks described below, to nuance between this
definition and the term "black-box" widely used in the literature.
2. The attacker’s goal : depends on whether he simply wants to deceive the system, or to induce a precise
prediction for certain inputs. Two forms of attack can be listed:
• Targeted attacks: direct the ML algorithm to a specific class, i.e., the adversary tricks the classi-
fier into predicting all adversary examples as a specific target class.
• Non-targeted attack: aims to misclassify the input sample away from its original class, regardless
of the new output class. They are easier to implement because more alternatives are available to
reorient the output. Note that in binary classification problems, targeted and untargeted attacks
are equivalent.

3.2. Survey of adversarial attack generation approaches


As many adversarial attack generation approaches can be applied to both targeted and untargeted scenar-
ios, we will rather rely on the attacker’s knowledge to classify them.

3.2.1. White-box attacks


Fast Gradient Sign Method (FGSM) [26]. creates adversarial examples by adding noise to the original
sample along the gradient directions.
Two iterative extension of FGSM, namely, Basic Iterative Method (BIM) [39] and Projected Gradient
Descent (PGD)[43] have been also used in the recent literature.

6
Figure 2: Adversarial attack techniques

Attack generation techniques

White-box attacks: the attacker Gray/black-box attacks: the attacker has no or


has full access to all informa- very little knowledge of the ML-based NIDS.
tion on the ML-based NIDS

Examples:
Score-based Decision-based Transfer-based
Fast Gradient Sign Method (FGSM) [26]
attacks: the attacks: the attacks: the
Basic Iterative Method (BIM) [39]
attacker has attacker has only attacker has access
Projected Gradient Descent (PGD)[43]
access to the access to the final to the training data-
The Jacobian-based Saliency
predicted scores model decision set and use to train
Map Attack (JSMA) [55]
(probability vector) (binary decision) a substitute model
Carlini and Wagner attack (C&W) [15]
DeepFool (DF) [48]
Examples: Examples:
Zeroth-order Boundary at-
optimization tacks [13]
(ZOO) attacks [18] Hop Skip Jump
Attack attack [17]

The Jacobian-based Saliency Map Attack (JSMA) [55]. generates adversarial examples using forward deriva-
tives (i.e., model Jacobian). JSMA iteratively perturbs features/components of the input one at a time instead
of perturbing the whole input to fool the classifier.

Universal Adversarial Perturbations (UAP) [47]. are a special type of untargeted attacks that consist on
creating a constant perturbation that successfuly misclassifies a specified fraction of the input samples.

DeepFool (DF) [48]. is an untargeted attack based on computing the minimum distance between the original
input and the decision boundary.

Figure 3: defence techniques

Defence techniques

Modify the input Modify the classifier Add an external model

Examples: Examples: Examples:


Gaussian data Gradient GAN based
augmentation [76] Masking [52] approaches [42]

7
Carlini and Wagner attack (C&W) [15]. The authors formulate the search for an adversarial sample as an
optimization problem with the following objective:

min D(x, x + ϵ) + c. f (x + ϵ) subject to x + ϵ ∈ D


ϵ

where ϵ denotes the adversarial perturbation, D(., .) denotes the ℓ0 , ℓ0 or ℓ∞ distance metric, and f (x + ϵ)
define the cost function such that f (x + ϵ) ⩾ 0 if and only if the model correctly classifies x + ϵ (i.e., gives it
the same label as x).

3.2.2. Gray/black-box attacks


Score-bases attacks. : the attacker has access to the predicted scores, i.e., the probability of each predicted
label to belong to the classification classes of the model. Examples include the Zeroth-order optimization
(ZOO) attacks [18].

Decision-based attacks. : the attacker only has access to the final decision of the model (binary decision)
without any confidence score. Examples include Boundary attacks [13] and Hop Skip Jump Attack [17].

Transfer-based attacks. : the attacker has access to the hole or part of the training data-set and use it to train
another fully observable model, called "a substitute model" intending to emulate the attacked model called
"target model". Adversarial perturbations that can be synthesized from the "substitute model" are used to
attack the "target model".
We refer the reader to [16, 60, 54, 71] for more additional information on adversarial attacks. Figure 2
summarizes the different Adversarial attack generation techniques described above.

3.3. Defence
A defence technique aims at improving the robustness of the model against adversarial attacks. In [4],
the three following categories of defence techniques are highlighted:
• Modify the input data: These techniques do not deal directly with training models, but rely on modify-
ing the training data during training or modifying the input data during testing. For example Gaussian
data augmentation [76] technique involves augmenting the original data-set with copies of the origi-
nal samples to which Gaussian noise has been added. The underlying idea is that forcing the model to
make the same prediction for a true instance and its slightly perturbed version should increase its gen-
eralization capabilities. This method is widely used because of its simplicity, ease of implementation
and effectiveness against both gray/back-box and white-box attacks.
• Modify the classifier: This involves modifying the original classification model by changing the loss
functions, adding additional layers/sub-networks, etc. For example, the Gradient Masking method
modifies a machine learning model to mask its gradient from an attacker.
• Add an external model: these methods keep the original model intact and add one or more external
models to it during testing. For example, the authors of [42] used Generative Adversarial Networks
(GAN) to train the network along a generator network that attempts to generate a perturbation to that
network.
Figure 3 summarize the different defence techniques described above.

8
3.4. Defence and attacks in IDS
Table 1 presents and compares recent research on ML based NIDS in adversarial environment. For
each research work, we highlight i) the evaluated ML classifiers ii) the evaluation data-set iii) the adversarial
attack algorithms and iv) the defence techniques, if any. In particular, we classify the evaluated ML classifiers
into two categories: shallow and deep learning. We also divide the adversarial attack generation techniques
into State of the art techniques, i.e. techniques inspired by the field of computer vision, and new techniques
designed by the authors.
The first row of the table, presents a statistic revealing the trends in the literature. It can be seen that the
majority of the literature (95%) evaluate the robustness of deep learning techniques, while a minority (37%)
evaluates shallow learning. The majority of the latter focus on a single type of adversarial attack, proposed
by the authors, and do not address the various adversarial attacks widely used in the literature.
The evaluation of shallow ML based NIDS under adversarial environment requires further study. To fill
this gap, this paper evaluates shallow ML based NIDS against the most used attack generation approaches in
the literature. Only [56] have already addressed this issue, but the authors did not explore defence techniques.
In more detail, we evaluate diverse and widely used ML algorithms [46] in the NIDS domain when ex-
posed to white-box and gray/back-box adversarial attacks generated by well know SOA algorithms. Further-
more, we explore the effect of the Gaussian data augmentation defence technique on different classification
settings.

4. Evaluating IDS robustness against adversarial attacks

The white-box attacks investigated in this paper are: FGSM attack, BIM, PGD attack, JSMA , Deep-
Fool attack, Carlini and Wagner attack. The gray/black-box examined attacks are: Zoo Attack, Boundary
attack and Hop Skip Jump Attack. Moreover, we generate additional adversary attacks in a naive way using
Gaussian noise, with different intensities (σ values are 0.01, 0.1 and 0.2).
In order to fairly compare the robustness of diverse classifiers, we have to apply the same attacks, with
the same configuration and the same hyper-parameters to all classifiers. However, white-box attacks are
highly dependent on the type of classifier they attack, e.g., FGSM, BIM, PGD, and JSMA use the gradient
of the classifier to generate the attacks, and thus can only be applied to gradient-based classifiers.
To overcome this problem, we propose to use an external DNN-based surrogate classifier, which we
call "Generator" to which we apply all white-box attacks in order to generate the adversary samples, under
the same conditions. The samples generated by each type of white-box attack are then introduced in the
classifiers in order to measure their robustness against such an attack. This idea is based on the transferable
property of adversarial attacks [21], which shows that the effect of the attack can be transferred to other ML
models, including the "Generator" in our case. We use a DNN composed of 7 completely connected layers
with dimensions ranging from 1024 to 32. From one layer to another, the dimension is divided by 2. The
architecture of the generator is different from the evaluated neural network. It is more complex to make the
generation of the adversarial samples more complex.
To improve the robustness of IDSs, defensive techniques can be applied. To obtain the most robust NIDS
system, the manufacturer must follow an iterative procedure:
1. Build a basic NIDS

2. Evaluate its robustness against a set of adversarial attacks


3. Apply a defence technique to increase its robustness

9
4. Repeat 2) and 3) until the level of robustness of the model is acceptable.
In this paper, we focus on steps 2) and 3), i.e. we evaluate the robustness of the classifiers against adversarial
attacks, then apply the Gaussian data augmentation defence technique and measure its contribution to
improving its robustness.
In the following, we describe the experimental setup, then present the results.

4.1. Data-sets description


4.1.1. NSL-KDD data-set [5]
The NSL-KDD data-set is derived from the KDDCup 99 data-set and addresses the problems of the
latter, namely irrelevant records and data imbalance between normal and abnormal records. A record is
defined by 41 features, including 9 basic features of individual TCP connections, 13 content features within
a connection, 9 temporal features calculated within a two second time window, and 10 other features. The
data-set contains 24 attack types, grouped into 4 categories of attacks, namely denial of service (DoS),
remote to local (R2L), user to root (U2R) and probing. The data-set is divided into training and test subsets
containing 100.778 and 25.195 samples respectively. This data-set is publicly available 1 . The original
training NSL-KDD Train+ and testing NSL-KDD Test+ sets of the NSL-KDD data-set are used in this study.

4.1.2. UNSW-NB15 [50]


The data-set was created by by the cyber security research group at the Australian Centre of Cyber
Security (ACCS) in 2015. It contains approximately 100 GB of raw data (normal and malicious traffic). 9
different types of attacks were used to generate the malicious traffic: Fuzzers, Analysis, Backdoors, DoS,
Exploits, Generic, Reconnaissance, Shellcode and Worms. Each sample is described by 49 features that
were generated by several feature extraction tools. The data-set is divided into training and test subsets
containing 175.341 and 82.332 samples respectively. The original training UNSW_NB15_training-set and
10 % (randomly selected samples) of the testing UNSW_NB15_testing-set are used in this study.

4.1.3. Data pre-processing


Furthermore, the training set is divided into training and validation sets according to 8:2. In order to
provide more suitable data for the classifier, One Hot Encoding and Data Normalization steps are performed.
One Hot Encoding involves encoding the categorical attributes, such as protocol, service, and state, into one-
hot numeric array. This data is then normalized between 0 and 1 to produce more homogeneous values.

4.2. The evaluated classifiers


We evaluate the performance of various well known ML algorithms: Adaboost, Bagging, Gradient boost-
ing, Logistic regression, Decision Tree, Random Forest, Support Vector Classifier (SVC) and also a Deep
Learning Network. Binary classification is considered. All our models were implemented using the Tensor-
Flow [1], Keras [28] and scikit-learn packages [51]. To generate adversarial samples, we use the open-source
IBM Robustness Toolbox (ART) framework [53]. All the hyper-parameters of the classifiers and algorithms
used to generate the adverse samples have been set to their default values to facilitate the comparison of the
different evaluation scenarios.

1 http://www.unb.ca/cic/datasets/nsl.html

10
(b) Applying defence techniques

(a) Initial robustness evaluation

Figure 4: Evaluation protocol

4.3. Robustness indicators


We consider the following two metrics:
• Accuracy: measures the ability of the IDS to correctly classify the malicious and legitimate traffic.
It represents the percentage of the number of correctly classified records out of the total number of
records. ACC = T P+TT N+FP+FN
P+T N

• False Negative Rate: is a more specific indicator that highlights the percentage of malicious traffic
that has successfully passed the IDS. FNR = FN+T
FN
P

Where TP are the True Positives and represent the number of anomalous records that are correctly identified
as anomalies. The TN are the True Negatives and calculate the number of normal records that are correctly
identified as normal. The FP are the False Positives that are the number of normal records that are mis-
classified as anomalous. The TN are the True Negatives and represent the number of anomalous records that
are identified as normal.
A good classifier is a classifier having high accuracy and a low False Negative Rate.

4.4. Evaluation protocol


4.4.0.1. Initial robustness evaluation. This scenario (see Fig. 4a) evaluates the performance of the classi-
fiers against adversarial attacks.

11
To generate adversarial samples using white-box attacks, the samples of test data-sets (NSL-KDD Test+
and UNSW_NB15_testing-set) are perturbed using gray/black-box attacks applied directly to the classifier
and white-box attacks applied to the "Generator" model. The generated adversarial samples form new test
data-sets, which we call, "Adversarial data-sets" (an adversarial data-set generated from NSL-KDD Test+
and another generated from UNSW_NB15_testing-set). The performance of the classifiers are evaluated in
the following scenarios:

• Scenario 1: measuring performance in baseline scenario


– Train: NSL-KDD Train+ / UNSW_NB15_training-set
– Test: NSL-KDD Test+ / UNSW_NB15_testing-set

• Scenario 2: measuring performance in adversarial environment


– Train: NSL-KDD Train+ / UNSW_NB15_training-set
– Test: Adversarial data-sets
The first sub-tables of Tables 4 and 5 in the appendix shows the results of accuracy, whereas the second
sub-tables shows the results for the False Negative Rate.
For each table, the first row shows the results of the first scenario whereas other rows illustrate the
results of the second scenario. The difference in performance between the baseline and adversarial scenarios
is highlighted in red for increase and yellow for decrease, for each classifier. For example, if a classifier’s
accuracy is 50% in the baseline scenario, if its performance decreases to 47% in the adversarial scenario, 3%
is highlighted in yellow, if its performance increases to 55% in the adversarial scenario, 5% is highlighted
in red.

4.4.0.2. Applying defence techniques. In this scenario (see Fig. 4b), we measure the contribution of Gaus-
sian data augmentation to the robustness improvement of NIDS classifiers. Therefore, two new data-set
called "augmented data-sets" are generated by applying Gaussian data augmentation on KDD Train+ and
UNSW_NB15_training-set. The performance of the classifiers is then evaluated in the following two sce-
narios:
• Scenario 3: measuring performance in non adversarial environment with training in augmented data
Train: Augmented data-sets, Test: NSL-KDD Test+ / UNSW_NB15_testing-set

• Scenario 4: measuring the impact of defence techniques in adversarial environment Train: Aug-
mented data-sets, Test: Adversarial data-sets
The obtained results are described in the third and fourth sub-tables of Tables 4 and 5, for accuracy and FNR
respectively. For each sub-table, the first row shows results of scenario 3, and the other rows illustrate the
results of scenario 4.

5. Experimental Results

5.1. Initial robustness evaluation


5.1.1. Scenario 1: measuring the performance in baseline scenario.
Results in the first raws of the first sub-tables show that almost all the classifiers perform well, the
accuracy varies between 74% and 77% on NSL-KDD except for Adaboost that has the worst performance

12
(accuracy 55.13% on NSL KDD and 60.6% on UNSW-NB15). DNN is the most efficient with an accuracy of
77.68% on NSL-KDD and 78.56% on UNSW-NB15. On the NSL-KDD data-set, these results are confirmed
by the False Negative Rates (first raw of the second sub-table). Indeed, Adaboost wrongly classifies 71.35%
of the malicious traffic as benign, while the other classifiers have a FNR varying between 29.68% and
39.32%. However, all classifiers have extremely low false negative rates (below 6%) on the UNSW-NB15
database.

5.1.2. Scenario 2: measuring the performance in adversarial environment.


Two analysis are of interest in this scenario: i) measuring the impact of each adversarial attack on
different classifiers and ii) measuring the performance of each classifier against different types of attacks.
The first analysis identifies the most/least vulnerable classifier to a given adversarial attack while the second
finds the most/least powerful adversarial attack for each classifier.

5.1.2.1. Accuracy. Gaussian noise attack:


• There are two different behaviors; some classifiers (DT, GB and bagging on UNSW-NB15 ) are very
vulnerable to the attack of Gaussian noise, even at low intensity (σ = 0.01), while the other classifiers
are more robust. In particular, for σ = 0.01, the performance of SVC (resp. Logistic regression)
remains stable on NSL-KDD, (resp. UNSW-NB15).
• For a high-intensity attack (σ = 0.2), the behavior is similar, where, on the NSL-KDD dataset, DT
loses almost half of its performance and GB drops to a third of its efficiency. However, the accuracy
decrease for the other algorithms ranges from 4.7 % for LR to 11.5% for Bagging. The results are
similar on the UNSW-NB15 database.
• Surprisingly, Adaboost, which has the worst baseline performance, is the least vulnerable to Gaussian
noise attacks on the NSL-KDD dataset, beating DT and GB when they all face this type of attack.
More surprisingly, a small Gaussian noise perturbation even increases Adaboost’s performance by
5.6%.
Gray/black-box attacks:
• Compared to the Gaussian noise attack, all classifiers are rather robust to the ZOO attack, with the
performance loss ranging from 0.3% for Adaboost to 5% for DNN on NSL-KDD. Thus, the DNN
with the best baseline performance is the most vulnerable to the ZOO attack.
• HopSkipJump and Boundary attacks have almost the same impact on all classifiers (whose perfor-
mance drops drastically). This can be explained by the fact that both attacks are of the same family,
i.e. HopSkipJump is an extension of the Boundary attack.
• With the exception of Adaboost, whose accuracy decreases by only 21% on NSL-KDD for both at-
tacks, the other classifiers are much more vulnerable. In fact, on NSL-KDD, DNN and RF lose 71%
and 68% of their performance respectively when dealing with the Boundary attack. As for the Hop-
SkipJUmp attack, both Bagging and DT lose almost 68% of their performance.
• The behavior of the majority of classifiers against these attacks is similar in both databases.
White-box attacks :
• FGSM, PGD and BIM which are attacks of the same family have almost the same impact on the
classifiers.

13
• JSMA is the most powerful attack; on NSL-KDD, the decrease in accuracy reaches 36% for GB.
C&W is slightly more powerful than the trio of FGSM, PGD, BIM, but these are all weak attacks that
result in a drop in accuracy of no more than 4.7%.
• Adaboost and Random Forest are the only classifiers that are robust to all white-box attacks. The
decrease in accuracy is limited to 1.4% and 3.7% for Adaboost and RF respectively on NSL-KDD.

• The classifiers are more vulnerable to these attacks on the UNSW-NB15 database than on the NSL-
KDD database.

5.1.2.2. False Negative Rate. The results are shown in the second sub-tables of Tables 4 and 5 . The
objective of this evaluation is to measure the ability of classifiers to block malicious traffic. Recall that the
FNR measures the percentage of malicious traffic classified as legitimate, so the lower the FNR, the better
the performance of the classifier.
Gaussian noise attack:

• The results show that FNR can decrease after data perturbation, i.e., malicious data initially classi-
fied as illegitimate are misclassified as legitimate after Gaussian noise perturbation. This reflects an
improvement in classifier performance, as is the case for Adaboost, GB, LR, and SVC on NSL-KDD
data-set. As for the Bagging, DT, DNN, and RF classifiers, the FNR increases by 23%, 13.2%, 22%,
and 11% for a σ = 0.02, reflecting the general decrease in accuracy described in the first sub-table of
Table.4.

• The GB result on NSL-KDD data-set is particularly interesting because the Gaussian noise has a
contradictory impact on the accuracy and the FNR. Indeed, the overall performance of the classifier
degraded (accuracy decreased by up to 21.68% for sigma =0.2), while the FNR also decreased (by
25.6% for sigma =0.2) which means that more malicious traffic was blocked. Thus, the degradation
in overall classifier performance may be due to mis-classifying benign traffic as illegitimate, which
causes False Alarms.
• We note that in the UNSW-NB15 database, where all classifiers had low false-negative rates in the
absence of adversarial attacks, the false-negative rates increase significantly, showing their sensitivity
to these attacks, especially for the Bagging and RF classifiers.

Gray/black-box attacks:

• Adaboost is robust to all three gray/back-box attacks on NSL-KDD data-set. Its FNR increase does not
exceed 3.1% (in the case of the HopSkipJump attack). However, on UNSW-NB15, the FNR increases
exponentially against Boundary and HopSkipJum and reaches 98.23%.
• HopSkipJump and Boundary attacks have the same impact on the other classifiers. Indeed, the FNR
reaches 100%, which means that the adverse noise manages to mis-classify all malicious samples into
benign ones.

• HopSkipJump and Boundary are more powerful than ZooAttack in most cases.
• Unlike NSL-KDD, some algorithms (DNN, LR,RF, SVC) are very robust to these attacks on the
UNSW-NB15 from the point of view of FNR, which does not increase.

White-box attacks:

14
• on NSL-KDD data-set, JSMA has the greatest influence on the deviation of FNR, whether it is positive
or negative. It is mainly noticed that the FNR of DT increases by 41% for DT and by 41% and 10%
for Adaboost and LG respectively.
• While the results are not significantly different from the baseline results for FGSM, PDG and BIM on
NSL-KDD data-set, we note that DNN is the most impacted classifier (an increase in FNR up to 8.7%),
while the FNR of Adaboost and Gradient Boosting decrease slightly, reflecting better classification of
malicious traffic.
• The impact of these attacks on the FNR of the classifiers is much more remarkable on the UNSW-
NB15 data-set, which increases exponentially for most of the classifiers except for Adabooost which
keeps a low FNR.

5.2. Applying defence techniques


5.2.1. Scenario 3: measuring the performance in non adversarial environment with training in augmented
data.
Comparing the performance of the classifiers trained on the augmented database in non adversarial envi-
ronment (first rows of Tables 4 and 5) with their performance when trained on initial training data-sets (first
row of the sub-tables 3), we notice that the performance of most of them (Bagging, DT, DNN, RF) has re-
mained stable on NSL-KDD data-set. The performance of Adaboost increased from 55.13% to 66.33%, but
the performance of LR and SVC deceased from 75% to 67.27% and from 74.29% to 65.36% respectively.
The FNR results (first row of Table 4) reflect the same conclusions. The performance degradation is more
noticeable in the UNSW-NB15 base.

5.2.2. Scenario 4: measuring the impact of defence techniques.


5.2.2.1. Accuracy. The results are shown in the third sub-tables and will be compared to the results of the
first sub-tables, where the accuracy was measured without applying defence techniques.
• Gaussian noise attack: The defence technique has improved the robustness of most classifiers. The
improvement is even more remarkable for a Gaussian perturbation of high intensity (σ = 0.2). The
improvement is almost perfect for SVC and LR that have become more robust but at the cost of
a degradation in performance, even in non adversarial conditions. More interesting, the DNN has
become more robust while keeping almost the same original performance on NSL-KDD data-set, it
is the classifier that has the best accuracy under normal and adversarial conditions on NSL-KDD
data-set.
• Gray/black-box attacks: The defence technique has different effects depending on the classifier. For
example on the NSL-KDD data-set: i) it contributes to the improvement of the robustness for some
classifiers as for LR and SVC whose accuracy decreased, compared to the baseline performance,
by only 55.3% and 53.8% (resp.) when confronted to a Boundary attack, versus 67% and 66.8%
(resp.) of decrease before having applied the defence technique. ii) it degrades the robustness of some
algorithms as it is the case for Adaboost whose accuracy decreases by 53% for a Boundary attack
instead of 22% without defence. iii) the defence technique has very little effect on the other classifiers
who have kept almost the same level of robustness as before. The behavior of the classifiers is similar
in both databases.
• White-box attacks: The defence technique is effective for almost all the classifiers against the FGSM,PGD
and BIM. As for JSMA, the robustness is also improved for all classifiers, and Gradient Boosting has
the best performance. The behavior of the classifiers is similar in both databases.

15
5.2.2.2. False Negative Rate. The results are shown in the fourth sub-table, and will be compared with the
second sub-tables, where the FNR is measured without defence technique.
• Gaussian Noise attack: The increase of the FNR is less important with the defence technique for most
of the algorithms thus their robustness has improved. However, the ability of these classifiers to block
malicious traffic has decreased overall, for example for SVC, the FNR is 50% while it did not exceed
37.52% without defence on NSL-KDD. Moreover, the defence technique did not improve the perfor-
mance Gradient Boosting and RF on NSL-KDD since their FNR increases from 10.2% to 48.51% and
from 45.07% to 55.51% respectively, thus more malicious traffic was blocked without defence. DNN
has the lowest FNR (the decrease in FNR does not exceed 2.8%) and a very good robustness. The de-
fence technique has improved its robustness.Curiously, the defence method decreased the robustness
of some algorithms to Gaussian noise attacks on UNSW-NB15 (FNR increases more) as is the case
for DT, Gradient Boost and Adaboost.
• Gray/black-box attacks: on NSL-KDD data-set, the defence has degraded the robustness of Adaboost
against Boundary and HopSkipJump (FNR increases by 49% while the increase was limited to 3.1%
before defence). It slightly improved the robustness of Bagging, DNN,RF). The improvement is more
significant for LR and SVC. FNR increased by only 52% and 49% for LR and SVC respectively,
compared to an increase by 62% for both without defence. On the UNSW-NB15 data-set, the defense
method proved to be effective and significantly improved the robustness of the classifiers in terms of
FNR.
• White-box attacks: After defence, the robustness of the classifiers for FGSM, PGD, BIM and C&W is
stable on NSL-KDD, as the classifiers are already quite robust. But the improvement of the robustness
is more visible for JSMA, although the global performance has degraded for most of the classifiers
(higher FNR, compared to Table 2). The defense method was not effective on UNSW-NB15 and even
degraded the performance of some classifiers (e.g. Gradient Boost, DT).

5.3. Analysis and Discussion


In this section we summarize and discuss our findings.

5.3.1. Global conclusions


• An attack does not impact all classifiers in the same way. Similarly, the robustness of a classifier
depends on the attacks. In the same sense, a defence technique is not effective against all attacks
and does not have the same effect on different classifiers (it can improve or decrease the classifier’s
robustness or be ineffective). Similarly, the behavior of a classifier when faced with an attack or a
defense method depends on the database.
• The robustness and overall performance of classifiers can be contradictory. As seen in the results on the
NSL-KDD data-set, Adaboost is a very robust classifier, but does not have high accuracy Ċonversely,
DNN is very efficient but is the most vulnerable to ZOO attack. Therefore, depending on the situation
and need of the IDS, robustness or performance can be privileged. Namely, if the IDS operates in
a certain environment, it is natural to favor performance, however, if the environment is uncertain,
robustness becomes important. Similarly, defence techniques can improve the robustness of classifiers
but at the cost of degrading their performance. Thus, a trade-off between these two objectives must
be considered. A good defence technique, improves the robustness of the classifier without degrading
its performance. In addition, the effectiveness of a defense method against a database attack can vary
from database to database.

16
5.3.2. Specific remarks
• Attacks of the same family (Boundary and HopSkipJump) have the same impact on each classifier.
• Sometimes, an attack can have an opposite effect: the Gaussian noise attack improved Adaboost’s
performance on NSL-KDD. In addition to mis-classifying more malicious traffic as legitimate, which
is the main objective of an adversarial IDS attack, an adversary attack can also increase the False
Alarms rate, as was the case with Gradient Boosting against Gaussian noise.
• Boundary and HopSkipJump attacks succeed in mis-classifying 100% of the malicious traffic on NSL-
KDD.

• The Gaussian data augmentation defence was especially effective on the Gaussian noise attack proba-
bly because they are of the same family.
• A defense method can even degrade the robustness of a classifier (e.g. DT against white box attacks
on the UNSW-NB15) , so it must be chosen appropriately.

It is also interesting to note that in our experiment, gray/back-box attacks are more effective than white-
box attacks, which is counter-intuitive since the latter have access to more information about the classifier.
This can be explained by the fact that gray/back-box attacks were applied directly on the classifiers while
white-box attacks were applied on a Generator network. On the other hand, the performance of the attacks
strongly depends on the setting of the hyper parameters. A better setting of the hyper-parameters would
probably enhance the performance of white-box attacks.

5.4. Comparison with the state of the art


As mentioned in 3.4, the paper [56] is the closest to our work since, like us, the authors evaluate the
robustness of shallow classifiers against state of the art adversarial attacks using the NSL-KDD database.
Since the hyper-parametric information is unavailable in [56], in addition to the difference in the list
of classifiers and attacks considered, we opt for qualitative comparison our evaluation framework with the
approach presented in [56].
The authors of [56] evaluate the robustness of 3 shallow classifiers (LR, RF,SVM) against three adversar-
ial white-box attacks (FGSM, PDG, L-BFGS) and a black-box attack SPSA, and do not propose a defence
method (Cf. Table 1). This paper, however, evaluates a richer and more varied collection of ML classi-
fiers (Adaboost, Bagging, DT, GB, LR, RF, SVC), against more varied and numerous adversarial attacks
(Gaussian noise, White-box attacks: FGSM,PGD,BIM,C&W, JSMA, Gray/black-box attacks: Zoo, Bound-
ary, HopSkipJump). Moreover, we complete our study by evaluating the impact of a defence technique
(Gaussian data augmentation) on the improvement of the robustness of the classifiers. This allowed us to
draw interesting conclusions about the trade-off between robustness and accuracy of classifiers, as discussed
above. Note, however, that the results obtained in [56] confirm our findings regarding the vulnerability of
ML classifiers to different attacks with varying degrees of sensitivity.

6. Discussion

In this work, changes on the input samples are made on the feature vector, however, an attacker does
not have access to the input feature vector of the ML algorithm to be able to modify it, but instead must
generate actual traffic that respects the features described by the feature vector generated by the adversarial
attack (FGSM, PGD, etc.). This traffic must also retain its original functionality (malicious or benign).

17
This transition from feature vector to actual instance is referred to in the literature as the "feature space" to
"problem space" transition [57], it is specific to certain data types, where unlike images, this transition is not
trivially reversible, and can be complex.
One way to facilitate this transition is to judiciously perform some modifications on the original traffic
sample, in order to obtain a sample with characteristics close to those of the perturbed sample (generated
adversary sample), without altering the primary function of the traffic. Examples of possible manipulations
include i) filling and fragmenting or duplicating protocol data units (PDUs, e.g. packet, segment, datagram,
etc.) to modify their volumetric characteristics (e.g. flow size, number of packets, etc.), ii) delaying the
transmission of PDUs, to act on their temporal characteristics (e.g. inter-arrival time of packets), iii) modi-
fying the values of some fields, etc. In order not to alter the main function of the traffic, the modifications
must be made only on the fields that do not have an impact on this function. To do this, the PDU manipu-
lation tools need to be explored and improved. Fortunately, there are already promising tools such as Scapy
[61] which is a packet manipulation program. Custom synthetic traffic generators [2] can also be explored.

7. Conclusion and future work

This paper focuses on the research area of adversarial machine learning. We study the robustness of vari-
ous widely used ML classifiers against adversarial examples in the context of network IDS. We consider both
gray/black-box and white-box attacks. A DNN-based external classifier has been used to generate white-box
based adversarial examples. In addition, we studied the impact of a defence technique based on Gaussian
data augmentation to improve the robustness of different NIDS. For the evaluation, we consider both the
accuracy and the false negative rate. The latter measures the percentage of malicious traffic that successfully
bypasses the NIDS. The NSL-KDD benchmark data-set was used for the evaluation. The results show that
attacks do not have the same impact on all classifiers and that the robustness of a classifier depends on the
attack. Similarly, a defence technique is not effective for all classifiers, nor against all attacks. Further-
more, a defence technique may improve the robustness of a classifier but degrade its overall performance,
so a trade-off between performance and robustness must be considered depending on the NIDS application
scenario.
In future work, we intend to generate more realistic adversarial attacks that project more easily into
the problem space. To do so, we will follow some recommendations found in the literature, [70, 65, 44],
namely i) restrict the space of features to be perturbed, i.e., avoid perturbing non-differentiable features so
that the transformation is reversible, and the features directly related to the functionality of the flow so as
not to impact it, ii) perform small amplitude perturbations and check that the values of the modified features
remain valid (domain constraints), and iii) analyze the consistency of the values taken by the correlated
features.

8. Acknowledgement

We thank the anonymous reviewers for their constructive comments and suggestions that helped us
improve the quality of this work considerably.

References

[1] Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G.,
Isard, M., et al. (2016). Tensorflow: A system for large-scale machine learning. In 12th {USENIX}
symposium on operating systems design and implementation ({OSDI} 16), pages 265–283.

18
[2] Adeleke, O. A., Bastin, N., and Gurkan, D. (2022). Network traffic generation: A survey and methodol-
ogy. ACM Computing Surveys (CSUR), 55(2):1–23.
[3] Aiken, J. and Scott-Hayward, S. (2019). Investigating adversarial attacks against network intrusion
detection systems in sdns. In 2019 IEEE Conference on Network Function Virtualization and Software
Defined Networks (NFV-SDN), pages 1–7.

[4] Akhtar, N. and Mian, A. (2018). Threat of adversarial attacks on deep learning in computer vision: A
survey. Ieee Access, 6:14410–14430.
[5] Aljawarneh, S., Aldwairi, M., and Yassein, M. B. (2018). Anomaly-based intrusion detection system
through feature selection analysis and building hybrid efficient model. Journal of Computational Science,
25:152–160.

[6] Anthi, E., Ahmad, S., Rana, O., Theodorakopoulos, G., and Burnap, P. (2018). Eclipseiot: A secure and
adaptive hub for the internet of things. Computers & Security, 78:477–490.
[7] Anthi, E., Williams, L., Javed, A., and Burnap, P. (2021). Hardening machine learning denial of service
(dos) defences against adversarial attacks in iot smart home networks. computers & security, 108:102352.

[8] Antonakakis, M., April, T., Bailey, M., Bernhard, M., Bursztein, E., Cochran, J., Durumeric, Z., Halder-
man, J. A., Invernizzi, L., Kallitsis, M., et al. (2017). Understanding the mirai botnet. In 26th {USENIX}
security symposium ({USENIX} Security 17), pages 1093–1110.
[9] Apruzzese, G., Andreolini, M., Marchetti, M., Venturi, A., and Colajanni, M. (2020). Deep reinforce-
ment adversarial learning against botnet evasion attacks. IEEE Transactions on Network and Service
Management, 17(4):1975–1987.
[10] Apruzzese, G., Colajanni, M., and Marchetti, M. (2019). Evaluating the effectiveness of adversarial
attacks against botnet detectors. In 2019 IEEE 18th International Symposium on Network Computing and
Applications (NCA), pages 1–8. IEEE.

[11] Arp, D., Spreitzenbarth, M., Hubner, M., Gascon, H., Rieck, K., and Siemens, C. (2014). Drebin:
Effective and explainable detection of android malware in your pocket. In Ndss, volume 14, pages 23–26.
[12] Beigi, E. B., Jazi, H. H., Stakhanova, N., and Ghorbani, A. A. (2014). Towards effective feature
selection in machine learning-based botnet detection approaches. In 2014 IEEE Conference on Commu-
nications and Network Security, pages 247–255. IEEE.

[13] Brendel, W., Rauber, J., and Bethge, M. (2018). Decision-based adversarial attacks: Reliable attacks
against black-box machine learning models. In International Conference on Learning Representations.
[14] Buczak, A. L. and Guven, E. (2015). A survey of data mining and machine learning methods for cyber
security intrusion detection. IEEE Communications surveys & tutorials, 18(2):1153–1176.

[15] Carlini, N. and Wagner, D. (2017). Towards evaluating the robustness of neural networks. In 2017 ieee
symposium on security and privacy (sp), pages 39–57. IEEE.
[16] Chakraborty, A., Alam, M., Dey, V., Chattopadhyay, A., and Mukhopadhyay, D. (2018). Adversarial
attacks and defences: A survey. CoRR, abs/1810.00069.

19
[17] Chen, J. and Jordan, M. I. (2019). Boundary attack++: Query-efficient decision-based adversarial
attack. CoRR, abs/1904.02144.
[18] Chen, P.-Y., Zhang, H., Sharma, Y., Yi, J., and Hsieh, C.-J. (2017). Zoo: Zeroth order optimization
based black-box attacks to deep neural networks without training substitute models. In Proceedings of
the 10th ACM workshop on artificial intelligence and security, pages 15–26.
[19] Clements, J., Yang, Y., Sharma, A., Hu, H., and Lao, Y. (2019). Rallying adversarial techniques against
deep learning for network security. arXiv preprint arXiv:1903.11688.
[20] Creech, G. and Hu, J. (2013). Generation of a new ids test dataset: Time to retire the kdd collection.
In 2013 IEEE Wireless Communications and Networking Conference (WCNC), pages 4487–4492. IEEE.
[21] Dong, Y., Pang, T., Su, H., and Zhu, J. (2019). Evading defenses to transferable adversarial examples
by translation-invariant attacks. In Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pages 4312–4321.
[22] Fu, X., Zhou, N., Jiao, L., Li, H., and Zhang, J. (2021). The robust deep learning–based schemes for
intrusion detection in internet of things environments. Annals of Telecommunications, 76(5):273–285.
[23] Gamage, S. and Samarabandu, J. (2020). Deep learning methods in network intrusion detection: A
survey and an objective comparison. Journal of Network and Computer Applications, 169:102767.
[24] Garcia, S., Grill, M., Stiborek, J., and Zunino, A. (2014). An empirical comparison of botnet detection
methods. computers & security, 45:100–123.
[25] Garcia, S., Parmisano, A., and Erquiaga, M. J. (2020). IoT-23: A labeled dataset with malicious and
benign IoT network traffic. More details here https://www.stratosphereips.org /datasets-iot23.
[26] Goodfellow, I., Shlens, J., and Szegedy, C. (2015). Explaining and harnessing adversarial examples.
In International Conference on Learning Representations.
[27] Guerra-Manzanares, A., Medina-Galindo, J., Bahsi, H., and Nõmm, S. (2020). Medbiot: Generation
of an iot botnet dataset in a medium-sized iot network. In ICISSP, pages 207–218.
[28] Gulli, A. and Pal, S. (2017). Deep learning with Keras. Packt Publishing Ltd.
[29] Han, D., Wang, Z., Zhong, Y., Chen, W., Yang, J., Lu, S., Shi, X., and Yin, X. (2020). Evaluating
and improving adversarial robustness of machine learning-based network intrusion detectors. arXiv:
Cryptography and Security.
[30] Han, D., Wang, Z., Zhong, Y., Chen, W., Yang, J., Lu, S., Shi, X., and Yin, X. (2021). Evaluating and
improving adversarial robustness of machine learning-based network intrusion detectors. IEEE Journal
on Selected Areas in Communications, 39(8):2632–2647.
[31] Hashemi, M. J., Cusack, G., and Keller, E. (2019). Towards evaluation of nidss in adversarial set-
ting. In Proceedings of the 3rd ACM CoNEXT Workshop on Big DAta, Machine Learning and Artificial
Intelligence for Data Communication Networks, pages 14–21.
[32] Ibitoye, O., Shafiq, M. O., and Matrawy, A. (2019). Analyzing adversarial attacks against deep learning
for intrusion detection in iot networks. In 2019 IEEE Global Communications Conference, GLOBECOM
2019, Waikoloa, HI, USA, December 9-13, 2019, pages 1–6. IEEE.

20
[33] Jeong, J., Kwon, S., Hong, M., Kwak, J., and Shon, T. (2020). Adversarial attack-based security
vulnerability verification using deep learning library for multimedia video surveillance. Multim. Tools
Appl., 79(23-24):16077–16091.
[34] Jiang, H., Lin, J., and Kang, H. (2022). Fgmd: A robust detector against adversarial attacks in the iot
network. Future Generation Computer Systems, 132:194–210.

[35] Kang, H., Ahn, D. H., Lee, G. M., Yoo, J. D., Park, K. H., and Kim, H. K. (2019). Iot network intrusion
dataset.
[36] Khamis, R. A. and Matrawy, A. (2020). Evaluation of adversarial training on different types of neural
networks in deep learning-based idss.

[37] Khamis, R. A., Shafiq, M. O., and Matrawy, A. (2020). Investigating resistance of deep learning-
based IDS against adversaries using min-max optimization. In 2020 IEEE International Conference on
Communications, ICC 2020, Dublin, Ireland, June 7-11, 2020, pages 1–7. IEEE.
[38] Koroniotis, N., Moustafa, N., Sitnikova, E., and Turnbull, B. (2019). Towards the development of
realistic botnet dataset in the internet of things for network forensic analytics: Bot-iot dataset. Future
Generation Computer Systems, 100:779–796.
[39] Kurakin, A., Goodfellow, I. J., and Bengio, S. (2017). Adversarial machine learning at scale.
[40] Labrotary, L. (1999). Darpa intrusion detection evaluation data set. Cambridge, MA: Massachusetts
Institute of technology. Retrieved January, 12:2009.

[41] LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998). Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11):2278–2324.
[42] Lee, H., Han, S., and Lee, J. (2017). Generative adversarial trainer: Defense to adversarial perturba-
tions with gan. arXiv preprint arXiv:1705.03387.

[43] Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and Vladu, A. (2018). Towards deep learning models
resistant to adversarial attacks. In 6th International Conference on Learning Representations, ICLR 2018,
Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings.
[44] Merzouk, M. A., Cuppens, F., Boulahia-Cuppens, N., and Yaich, R. (2022). Investigating the practical-
ity of adversarial evasion attacks on network intrusion detection. Annals of Telecommunications, pages
1–13.
[45] Mirsky, Y., Doitshman, T., Elovici, Y., and Shabtai, A. (2018). Kitsune: an ensemble of autoencoders
for online network intrusion detection. arXiv preprint arXiv:1802.09089.
[46] Mishra, P., Varadharajan, V., Tupakula, U., and Pilli, E. S. (2018). A detailed investigation and analysis
of using machine learning techniques for intrusion detection. IEEE Communications Surveys & Tutorials,
21(1):686–728.
[47] Moosavi-Dezfooli, S., Fawzi, A., Fawzi, O., and Frossard, P. (2017). Universal adversarial perturba-
tions. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 86–94.

21
[48] Moosavi-Dezfooli, S., Fawzi, A., and Frossard, P. (2016). Deepfool: A simple and accurate method
to fool deep neural networks. In 2016 IEEE Conference on Computer Vision and Pattern Recognition
(CVPR), pages 2574–2582.
[49] Moustafa, N., Hu, J., and Slay, J. (2019). A holistic review of network anomaly detection systems: A
comprehensive survey. Journal of Network and Computer Applications, 128:33 – 55.

[50] Moustafa, N. and Slay, J. (2015). Unsw-nb15: a comprehensive data set for network intrusion detec-
tion systems (unsw-nb15 network data set). In 2015 military communications and information systems
conference (MilCIS), pages 1–6. IEEE.
[51] Müller, A. C. and Guido, S. (2016). Introduction to machine learning with Python: a guide for data
scientists. " O’Reilly Media, Inc.".
[52] Nguyen, L., Wang, S., and Sinha, A. (2018). A learning and masking approach to secure learning. In
International Conference on Decision and Game Theory for Security, pages 453–464. Springer.
[53] Nicolae, M., Sinn, M., Minh, T. N., Rawat, A., Wistuba, M., Zantedeschi, V., Molloy, I. M., and
Edwards, B. (2018). Adversarial robustness toolbox v0.2.2. CoRR, abs/1807.01069.

[54] Ozdag, M. (2018). Adversarial attacks and defenses against deep neural networks: a survey. Procedia
Computer Science, 140:152–161.
[55] Papernot, N., McDaniel, P., Jha, S., Fredrikson, M., Celik, Z. B., and Swami, A. (2016). The limitations
of deep learning in adversarial settings. In 2016 IEEE European Symposium on Security and Privacy
(EuroS P), pages 372–387.
[56] Peng, Y., Su, J., Shi, X., and Zhao, B. (2019). Evaluating deep learning based network intrusion
detection system in adversarial environment. In 2019 IEEE 9th International Conference on Electronics
Information and Emergency Communication (ICEIEC), pages 61–66.
[57] Pierazzi, F., Pendlebury, F., Cortellazzi, J., and Cavallaro, L. (2020). Intriguing properties of adversarial
ml attacks in the problem space. In 2020 IEEE symposium on security and privacy (SP), pages 1332–
1349. IEEE.
[58] Qiu, H., Dong, T., Zhang, T., Lu, J., Memmi, G., and Qiu, M. (2020). Adversarial attacks against
network intrusion detection in iot systems. IEEE Internet of Things Journal, pages 1–1.

[59] Qureshi, A. U. H., Larijani, H., Yousefi, M., Adeel, A., and Mtetwa, N. (2020). An adversarial approach
for intrusion detection systems using jacobian saliency map attacks (jsma) algorithm. Computers, 9(3).
[60] Ren, K., Zheng, T., Qin, Z., and Liu, X. (2020). Adversarial attacks and defenses in deep learning.
Engineering, 6(3):346 – 360.
[61] Rohith, R., Moharir, M., Shobha, G., et al. (2018). Scapy-a powerful interactive packet manipulation
program. In 2018 international conference on networking, embedded and wireless systems (ICNEWS),
pages 1–5. IEEE.
[62] Rosenberg, I., Shabtai, A., Elovici, Y., and Rokach, L. (2020). Adversarial learning in the cyber
security domain. arXiv preprint arXiv:2007.02407.

22
[63] Sharafaldin, I., Lashkari, A. H., and Ghorbani, A. A. (2018a). Toward generating a new intrusion
detection dataset and intrusion traffic characterization. In ICISSp, pages 108–116.
[64] Sharafaldin, I., Lashkari, A. H., and Ghorbani, A. A. (2018b). Toward generating a new intrusion
detection dataset and intrusion traffic characterization. ICISSp, 1:108–116.

[65] Sheatsley, R., Papernot, N., Weisman, M. J., Verma, G., and McDaniel, P. (2022). Adversarial examples
for network intrusion detection systems. Journal of Computer Security, (Preprint):1–26.
[66] Sultana, N., Chilamkurti, N., Peng, W., and Alhadad, R. (2019). Survey on sdn based network intru-
sion detection system using machine learning approaches. Peer-to-Peer Networking and Applications,
12(2):493–501.

[67] Venturi, A., Apruzzese, G., Andreolini, M., Colajanni, M., and Marchetti, M. (2021). Drelab - deep
reinforcement learning adversarial botnet: A benchmark dataset for adversarial attacks against botnet
intrusion detection systems. Data in Brief, 34:106631.
[68] Vitorino, J., Oliveira, N., and Praça, I. (2022). Adaptative perturbation patterns: Realistic adversarial
learning for robust intrusion detection. Future Internet, 14(4):108.

[69] Wang, J., Pan, J., AlQerm, I., and Liu, Y. (2021). Def-ids: An ensemble defense mechanism against
adversarial attacks for deep learning-based network intrusion detection. In 2021 International Conference
on Computer Communications and Networks (ICCCN), pages 1–9. IEEE.
[70] Wang, N., Chen, Y., Xiao, Y., Hu, Y., Lou, W., and Hou, T. (2022). Manda: On adversarial example
detection for network intrusion detection system. IEEE Transactions on Dependable and Secure Com-
puting.
[71] Wang, X., Li, J., Kuang, X., Tan, Y.-a., and Li, J. (2019). The security of machine learning in an
adversarial setting: A survey. Journal of Parallel and Distributed Computing, 130:12–23.
[72] Wang, Z. (2018). Deep learning-based intrusion detection with adversaries. IEEE Access, 6:38367–
38384.
[73] Xu, H., Ma, Y., Liu, H., Deb, D., Liu, H., Tang, J., and Jain, A. (2020). Adversarial attacks and defenses
in images, graphs and text: A review. International Journal of Automation and Computing, 17:151–178.
[74] Xu, Y., Zhou, Y., Sekula, P., and Ding, L. (2021). Machine learning in construction: From shallow to
deep learning. Developments in the Built Environment, 6:100045.
[75] Yang, K., Liu, J., Zhang, C., and Fang, Y. (2018). Adversarial examples against the deep learning
based network intrusion detection systems. In MILCOM 2018 - 2018 IEEE Military Communications
Conference (MILCOM), pages 559–564.
[76] Zantedeschi, V., Nicolae, M.-I., and Rawat, A. (2017). Efficient defenses against adversarial attacks. In
Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security, AISec ’17, pages 39–49,
New York, NY, USA. ACM.
[77] Zenati, H., Foo, C. S., Lecouat, B., Manek, G., and Chandrasekhar, V. R. (2018). Efficient gan-based
anomaly detection. arXiv preprint arXiv:1802.06222.

23
[78] Zhang, C., Costa-Pérez, X., and Patras, P. (2020a). Tiki-taka: Attacking and defending deep learning-
based intrusion detection systems. In Proceedings of the 2020 ACM SIGSAC Conference on Cloud Com-
puting Security Workshop, pages 27–39.
[79] Zhang, C., Costa-Pérez, X., and Patras, P. (2022). Adversarial attacks against deep learning-based
network intrusion detection systems and defense mechanisms. IEEE/ACM Transactions on Networking.

[80] Zhang, S., Xie, X., and Xu, Y. (2020b). A brute-force black-box method to attack machine learning-
based systems in cybersecurity. IEEE Access, 8:128250–128263.
[81] Zong, B., Song, Q., Min, M. R., Cheng, W., Lumezanu, C., Cho, D., and Chen, H. (2018). Deep
autoencoding gaussian mixture model for unsupervised anomaly detection. In International Conference
on Learning Representations.

Appendices

24
Table 1: Summary of research works related to Document Alignment

Algorithm Adversarial attack Defence technique


Classic Deep
Work Year Data-set SOA Designed
ML Learning
(%) 37% 95% 53% 47% 53%
ZOO,Gan based
[75] 2018 ✗ DNN NS-LKDD ✗ ✗
attack
FGSM,JSMA
[72] 2018 ✗ DNN NSL-KDD ✗ ✗
DeepFool,C&W
AdaBoost
CTU [24]
DT,GB Alter some Remove
[10] 2019 MLP CICIDS[63] ✗
KNN, LR features values altered features
BOTNET [12]
RF,SVM
LR,RF FGSM,PGD
[56] 2019 DNN NSL-KDD ✗ ✗
SVM L-BFGS,SPSA
FGSM
JSMA
[19] 2019 ✗ KitNET [45] Kitsune [45] ✗ ✗
C&W
Elastic Net Method
FGSM
[32] 2019 ✗ FFNN,SNNs BoT-IoT [38] ✗ Feature Normalization
BIM, PGD
KitNET [45] Alter the
[31] 2019 Reconstruction
DAGMM [81] ✗ CICIDS [63] ✗ packets to
[58] 2020 from Partial Observation
BiGAN [77] mimic benign traffic
DARPA SYN
KNN,LR Perturb some
[3] 2020 ✗ flood [40] ✗ ✗
RF,SVM important features
CICIDS [63]
DT,LR ADFA-LD [20] GAN based attack
[80] 2020 MPL ✗ ✗
NB,RF DREBIN [11] Brute-force attack
Random Mutation Adversarial training
DT,IF Kitnet [45]
[29] 2020 Kitsune [45] ✗ and duplication Feature selection
LR,SVM MLP
of selected features Adversarial feature reduction
[33] 2020 ✗ CNN, AE MNIST [41] FGSM, JSMA ✗ ✗
CNN, RNN FGSM, BIM, C&W
[36] 2020 ✗ UNSW-NB15 ✗ Min-max
ANN PGD, DeepFool
[37] 2020 ✗ DNN UNSW-NB15 FGSM, BGA, BCA ✗ Min-max
Use saliency map
[58] 2020 ✗ KitNET [45] MIRAI [8] ✗ to identify critical ✗
features and perturb them
[59] 2020 ✗ RNN NSL-KDD JSMA ✗ ✗
Model Voting Ensembling
MLP, CNN
[78] 2020 ✗ CICIDS [63] JSMA ✗ Ensemble Adversarial Training
CNN, LSTM
Adversarial Query Detection
[9] 2020 Wide CTU [24] Deep RL Deep RL based
RF ✗
[67] 2021 and Deep BOTNET [12] attacks adversarial training
J48 DT, RF Altering some
[30] 2021 ✗ EclipseIoT [6] ✗ ✓
NB, SVM selected features
[7] 2021 LSTM, CNN, GRU ✗ CSE-CIC-IDS2018 [64] FGSM ✗ ✓
[22] 2021 ✗ LSTM, CNN, GRU CSE-CIC-IDS2018 [64] FGSM ✗ ✓
FGSM, BIM
[69] 2021 ✗ DNN CSE-CIC-IDS2018 [64] ✗ ✓
JSMA, DeepFool
Nes, Boundary
[79] 2022 ✗ MPL, LSTM, CNN CSE-CIC-IDS2018 [64] HopSkipJump, ✗ ✓
pointwise, Opt-attack
Adaptative
[68] 2022 RF MLP CIC-IDS2017 [64], IoT-23 [25] ✗ perturbation ✗
pattern method
Altering
[34] 2022 ✗ LSTM, RNN MedBIoT [27] and IoTID [35] dataset) [35]. ✗ ✗
features
Adaboost
GN,ZOO,BIM
Bagging
This NSL-KDD FGSM, JSMA, Gaussian data
- DT, GB DNN ✗
paper UNSW-NB15 C&W,PGD augmentation
LR, RF
BA, HSJ
SVC

25
Table 2: Summary of results for evaluation scenario 2

Behavior against a Gaussian noise attack


• DT and GB are very vulnerable compared to other classifiers
• For GB the attacks have caused an increase in False alarms.
• Adaboost is the most robust although it has the worst performance in the baseline scenario.
• The classifiers classify more malicious traffic as legitimate on UNSW-NB15 than on NSL-KDD (high false negative rates),
especially for Bagging and Random Forest.

Behavior against Gray/black-box attacks


• The performance of the classifiers drops drastically when faced to HopSkipJump and Boundary attacks
• Classifiers are rather robust to ZOO attack.
• DNN with the best baseline performance, is the most vulnerable to the ZOO attack.
• Adaboost is robust to three gray/back-box attacks on the NSL-KDD but extremely sensitive to Boundary and HopSkipJum
attacks on the UNSW-NB15.
• From the point of view of accuracy, the classifiers have the same behavior against attacks in both databases. However, the
behavior of FNR is different from one database to the other: it increases for most of the classifiers on the NSL-KDD database,
but remains very low for some classifiers (DNN, LR, RF SVC) on the UNSW-NB15 database.

Behavior against white-box attacks


• Classifiers are rather robust to FGSM, PGD, BIM and C&W attack also has a very similar effect.
• JSMA is the most powerful attack.
• Adaboost and RF are the only classifiers that are robust to all white box attacks.
• The classifiers were more vulnerable to these attacks on the UNSW-NB15 database than on the NSL-KDD database, for bothe
the accuracy and False Negative Rate.

26
Table 3: Summary of results for evaluation scenario 4

Behavior against a Gaussian noise attack


• The defence technique has improved the robustness of most classifiers.
• The robustness of SVC and LR has improved but at the cost of performance degradation.
• The robustness of the DNN has been improved on NSL-KDD dataset while maintaining its good performance, thus DNN has
the best accuracy under normal and adversarial conditions.
• The defence method decreased the robustness of some algorithms on the UNSW-NB15 (FNR increases more) as is the case
for DT, Gradient Boost and Adaboost.

Behavior against Gray/black-box attacks


• The robustness of Bagging, DT, LR,SVC has improved on NSL-KDD dataset
• The robustness of Adaboost has decreased on NSL-KDD dataset
• The defence technique has almost no effect on other classifiers. In terms of FNR, the defense method was more effective on
the UNSW-NB15 basis, than NSL KDD (FNR rates are lower).

Behavior against white-box attacks


• The robustness of classifiers against FGSM, PGD, BIM attacks is almost the same since they are already quite robust.
• Robustness against JSMA has improved for most classifiers but at the cost of performance degradation for most of them. In
terms of FNR, the defense method was not effective on UNSW-NB15 and even degraded the performance of some classifiers
(e.g. Gradient Boost, DT).

27
Table 4: Evaluation Results NSL-KDD

Attack Adaboost Bagging DTree DNN Grad. Boos. Logistic Regression Random Forest SVC
Table 1 Accuracy (Train: NSL-KDD Train+ , Test: NSL-KDD Test+ ) Vs (Train: NSL-KDD Train+ , Test: Adversarial data-set)

Table 2 False negative rate (Train: NSL-KDD Train+ , Test: NSL-KDD Test+ ) Vs (Train: NSL-KDD Train+ , Test: Adversarial data-set)

Table 3 Accuracy (Train: Augmented data-set, Test: NSL-KDD Test+ ) Vs (Train: Augmented data-set, Test: Adversarial data-set)

Table 3 False negative rate (Train: Augmented data-set, Test: NSL-KDD Test+ ) Vs (Train: Augmented data-set, Test: Adversarial data-set)

28

Legend:
Table 5: Evaluation Results UNSW-NB15

Attack Adaboost Bagging DTree DNN Grad. Boos. Logistic Regression Random Forest SVC
Table 1 Accuracy (Train: NSL-KDD Train+ , Test: NSL-KDD Test+ ) Vs (Train: NSL-KDD Train+ , Test: Adversarial data-set)

Table 2 False negative rate (Train: NSL-KDD Train+ , Test: NSL-KDD Test+ ) Vs (Train: NSL-KDD Train+ , Test: Adversarial data-set)

Table 3 Accuracy (Train: Augmented data-set, Test: NSL-KDD Test+ ) Vs (Train: Augmented data-set, Test: Adversarial data-set)

Table 3 False negative rate (Train: Augmented data-set, Test: NSL-KDD Test+ ) Vs (Train: Augmented data-set, Test: Adversarial data-set)

29

Legend:

You might also like