investigate the usage of the domain-exclusive
model. For generating an adversarial sample, a
phishing instance’s values can be modified to any
value that has already appeared in another phishing
instance. This approach is not applicable to textual
features such as the URL. For phishing instances
that are correctly classified, all possible adversarial
samples are generated. The authors computed the
adversary cost with a tuple of two values: the
number of features modified and the Euclidean
distance between the phishing instance and the
adversarial sample.
The experiments by Shirazi et al. were on four
published phishing datasets. The first dataset
includes 1000 legitimate websites instances from
Alexa [7] and 1200 phishing instances from
PhishTank [8]. The dataset contained eight features
which are: domain length, presence of a non-
alphabetic character in the domain name, the ratio of
hyperlinks referring to the domain name, the
presence of HTTPS protocol, matching domain
name with copyright logo, and matching domain
name with the page title. The second dataset is by
Rami et al. [9]. The dataset includes 4898 legitimate
instances from Alexa and 6158 phishing instances
from PhishTank. The dataset includes 30 features,
which are from five categories: URL-based,
abnormal-based, HTML-based, JavaScript-based,
and domain-based features. The third dataset is by
Abdelhamid et al. [10], which contains 1350
instances and 16 features. The fourth dataset is by
Tan et al. [11], which had 5000 instances for
legitimate and phishing, collected from Alexa and
PhishTank. The dataset included 48 features that
were extracted from the URL and HTML.
In our previous publication [2], we have
proposed PUCNN, a URL-only phishing model that
is based on a character level CNN model. For
training and evaluation, we have collected and
preprocessed MUPD (Massive URL Phishing
Detection) dataset which contained 1,167,201
phishing URLs and 1,140,599 legitimate URLs. The
source of phishing URLs was PhishTank, whereas
the legitimate URLs were collected from DomCop
top 10 million domains [12]. We have split MUPD
dataset into training, validation, and testing datasets
of the following proportions: 0.6, 0.2, and 0.2.
PUCNN achieved 95.78% accuracy in the testing
dataset. PUCNN outperformed RandomForestNLP
[13], a state-of-art URL-only model, in their
published dataset.
Wang et al. [14] proposed PDRCNN, a URL-
only phishing detection model which is also based
on a character level CNN. They trained and
evaluated their model on a dataset they collected
245,385 phishing URLs from PhishTank and
245,023 legitimate URLs from Alexa top 1 million.
Their model achieves 95.61% accuracy using 10-
fold cross-validation. However, the authors used
only CANTINA+ [15] as their only benchmark. The
main problem is that CANTINA+ is not a state-of-
art model. Additionally, CANTINA+ is a non-URL
exclusive model. Although, the authors retrieved old
phishing pages to train and evaluate CANTINA+,
we believe that it is likely that many of these pages
no longer represent the phishing pages as they were
reported a long time before collection, which makes
the benchmark less useful. PDRCNN is similarly a
textual URL model and can be evaluated under the
proposed adversarial scenario.
Furthermore, there exist many character-level
CNN architectures in the literature, such as those in
[16] and [17]. These CNN models have already
achieved excellent results in various text
classification and language modeling tasks.
Similarly, these CNN models can be applied to the
problem of phishing URL detection, and it is
possible to evaluate them under the proposed
adversarial scenario.
3 Methodology
In this section, we discuss the threat model and how
we simulate the attacker’s behavior by adversarial
sampling.
3.1 Threat Model
In this subsection, we discuss the attacker’s goal,
knowledge, influence, control, and constraints. It is
important to specify the threat model, as we attempt
to simulate it. Additionally, we show how and why
our threat model differs from the threat model used
by Shirazi et al. [6], which we discussed in related
work. Table 1 provides a comparison summary. Our
main motivation for having a different threat model
is that the attacker in Shirazi et al. threat model is
powerful, which can be seen from the results where
they found that they can reduce the recall to 0 by
controlling only four attributes. This threat model
reflects the adversarial scenario we discussed in
section 1.
3.1.1 Attacker’s Goal
In our threat model, we assume that the attacker
wants to attack the recall of the model. The attacker
seeks the generated adversarial samples to pass as
legitimate while they are phishing. In practice,
achieving this goal means that the attacker manages
to send the phishing website to the user, avoiding
WSEAS TRANSACTIONS on COMPUTER RESEARCH
DOI: 10.37394/232018.2022.10.1