would be preferable to work with 8 features, due to
the twice faster classification. These results led to
the conclusion that further testing with classifiers,
capable of discriminating all 10 types of attacks
should be done with a SVM, using RBF kernel. All
observations from above are also supported by the
evaluating parameters from Tables 5-8.
Classifying all 10 DDoS attacks (indexes from 1
to 10) and recognizing non-attack samples in the
same time (index 0) took the RBF SVM almost
twice longer, using 8 features rather than 10, during
training (Table 3). In the same time, 8 feature lead
to more than twice faster validation with
comparison to the 10 feature implementation. The
most accurately spotted attack is DoS UDP flood
(Table 9), followed by the DDoS UDP flood, DoS
TCP flood, Service Scan, and so on. The most hard
to discover attacks are the Data Exfiltration and
DDoS HTTP flood. The most probable reason for
this is the considerably smaller number of samples
for these attacks, present in the training set,
compared to the amount of samples for the rest
types of attacks. Nevertheless, the proportion of data
exchanged during the various tested attacks
corresponds to real-world scenarios and the
observed dependency should be taken as inherited
peculiarity of the single SVM classifier itself.
Obviously, to get as close as possible detection rate
for these rarely spot types of attacks, one possible
direction for future work it would be to construct a
cascade of classifiers. The variation between the
number of discovered attacks between the phases of
training and testing is negligible. When using 8
features, differences in detection accuracy for some
of the attacks, compared to that for 10 features, goes
as high as 3 times, as it is in the case of OS
Fingerprint, or around 40% for the non-attack
samples (Table 9).
The most mismatched non-attack samples, using
10 features (Fig. 4 a), are recognized as DoS TCP
flood (25.2%), the DoS TCP attack – with DDoS
TCP flood (3.0%), DoS UDP flood – with DDoS
UDP flood (1.0%), DoS HTTP flood – with DDoS
TCP flood (45.2%), which is with 81.6% higher
than the correctly found samples, DDoS TCP flood
– with DoS TCP flood (7.5%), DDoS UDP flood –
with DoS UDP flood (2.9%), DDoS HTTP flood –
with DDoS TCP flood (62.1%), close to 60% higher
than the number of the correctly recognized samples
for this particular attack, Keylogging – with Service
Scan (57.1%), OS Fingerprint – with Service Scan
(71.2%), again serious mismatch rate, and Service
Scan – with DDoS TCP (2.0%). All this ratios could
be observed from the confusion matrix after
classification over the test set, representing the
proportion of the classified samples by attack from
the actual number of samples for the same attack, as
shown in Fig. 4 – for 10 features in Fig. 4 a and for
8 features – in Fig. 4. b. Using 8 features, lead to
increase of the proportion of mismatches with closes
incorrect type of attack, as follows: twice for non-
attack samples, 9 times for the DoS TCP flood, 1.2
times for the DoS UDP flood, 1.08 times for the
DoS HTTP flood, 1.5 times decrease for the DDoS
TCP flood, 1.7 times for the DDoS UDP flood, 1.2
times decrease for the DDoS HTTP flood, 57.1%
decrease for the Keylogging, and 1.7 times decrease
– for the Service Scan (Fig. 4 b). Apart from
worsening of the classification accuracy for some of
the attacks, such as the DoS TCP flood or the non-
attack samples, there is also a positive trend for
other types of attacks, such as the Keylogging.
Decrease of the information redundancy in the
training set at 8 features, compared to 10, obviously
preserves better some of the relations for attacks,
which have smaller intensity as per the exchanged
data over the network, such as the Keylogging. It
would be practical to use this feature set, although
considerably more inaccurate for attacks with high
intensity of the generated traffic, for some more rare
activities, when specifically searching for them in a
monitored network. All these results are also
supported by the evaluating parameters, shown in
Tables 10-13. For some of the attacks with really
small number of instances in the training and the
testing set, some of the parameters are hard to
calculate, as the denominator of the equations for
them, tends to be very small, almost equal to 0, so
they are marked I the tables with N/A.
At the end of the discussion section, we make a
comparison with another implementation of a binary
SVM classifier (detector) of DDoS attacks,
proposed by other authors in [11]. It is tested over
the same dataset with the same 10 features as in this
study and it has the cost parameter being put to C =
1, using a Linear kernel, and having a training time
limit of 100 000 iterations. The confusion matrices
for this classifier and the best of our binary SVM
classifiers (10 features, RBF kernel, 100 000
iterations limit, C = 1) are shown in Table 14 as
proportion of the detected samples to all actual of
that type ones, given in %.
WSEAS TRANSACTIONS on INFORMATION SCIENCE and APPLICATIONS
DOI: 10.37394/23209.2022.19.1
Vanya Ivanova,
Tasho Tashev, Ivo Draganov