conversations can lead fraudsters on for weeks. Edwards et
al. [14] described some of the persuasive techniques used
by human scam-baiters to achieve these results, noting
that they often mirrored the tactics used by the scammers
they targeted. It is the automatic deployment of these
techniques against cybercriminals that Canham & Tuthill
advocate [6], and that we explore.
In this paper, we describe our implementation of
a mailserver that can engage in scam-baiting activi-
ties automatically. We develop three alternative response-
generation strategies using publicly-available email cor-
pora as inputs to deep learning models, and perform a
one-month comparative evaluation and 12-day concurrent-
engagement experiment in which our system interacted
with real fraudsters in the wild. In short, our contributions
are:
•We demonstrate that automated scam-baiting is
possible, with 15% of our replies to known scam
emails attracting a response and 42% of conversa-
tions primarily involving a human fraudster. Some
conversations lasted up to 21 days.
•Further, we compare different approaches to au-
tomated scam-baiting in a naturalistic experi-
ment using randomised assignment. We find that
human-designed lures work best at attracting
scammer responses, but text generation methods
informed by the methods of human scam-baiters
are more effective at prolonging conversations.
•We also engage the same group of scammers with
two identical scam-baiting instances simultane-
ously. We find that two concurrent instances at-
tracted responses from 25% of the targeted scam-
mers and 29% of these scammers engaged with
both scam-baiting instances simultaneously.
•We release our code as a platform which can
be deployed to test alternative response strategies
and iterate on our findings. We also release both
full transcripts of our automated system’s con-
versations and a collection of human scam-baiter
conversations, to guide the development of new
active countermeasures and provide insight into
scammer operations.
The rest of this paper proceeds as follows. In Section 2
we provide some background on scam-baiting as a human
activity, as well describing the fundamental models used
within our work. Section 3 outlines our deployment plat-
form. Section 4 describes the different corpora we make
use of for finetuning and prepatory evaluation. Section 5
describes said finetuning and classifier evaluation, while
Section 6 describes our main experiments, including the
results from our comparison of the different response
strategies. Section 7 reflects on our findings, their limi-
tations, and our suggestions for future improvements, as
well as considering misuse concerns. We conclude with a
summary of our key results and recommendations.
2. Background
To provide the essential basics for active scamming
defense, this section describes the significance of scam-
baiting activities and recent advances in text generation
that enable adaptive conversational AI.
2.1. Scam-baiting protects potential scam victims
Scam-baiting is a kind of vigilante activity, in which
scam-baiters reply to the solicitation emails sent by scam-
mers and enter into conversation with them, in order to
waste scammers’ time and prevent them from scamming
other potential victims. This activity has become an In-
ternet subculture with various scam-baiter communities
across the Internet. Past research on scam-baiting has
explored the various motivations of scam-baiters [53],
[62], the strategies they use in conversations [13], [14]
and the ethics of their activities [51].
We attach importance to scam-baiting activity because,
by wasting scammers’ time, scam-baiting can help to
protect other vulnerable people from being scammed.
Herley [22] argued that scam-baiting activity can sharply
reduce the number of victims found by scammers by
decreasing the density of viable targets (i.e., the targets
that can lead to financial gain), making them less likely
to harm the potential victims. This argument seems to
be upheld in practice, as well. Scam-baiting exchanges
generally end in frustrated invectives from scammers
once they have understood what is taking place [14],
and prominent scam-baiter and comedian James Veitch
reports pointedly about scammers pleading with him to
stop emailing them [6], [56]. Some scam-baiters also use
their activities as a means to prod scammers into reflecting
on what they are doing (e.g., [47]), but the effectiveness
of this last technique is unknown.
There are existing industrial email conversation prod-
ucts that make use of NLP techniques to craft re-
sponses [33], [39], [46]. These products may conceivably
be adapted for automatic scam-baiting; however, their
focus is providing automatic customer service, and they
are tested only on responding appropriately to business
email communications. As a result, the performance of
these products at scam-baiting—which may involve inten-
tionally drawing out conversations to better waste scam-
mer resources–has not been evaluated. Previous anti-spam
conversation systems, such as RE:Scam [60], have demon-
strated the basic viability of automatic responses for dis-
rupting scammer operations, using random selection from
a series of canned template responses. This is a promising
start, but we suspect that an automatic scam-baiter that
can respond to the content of a scam message could be
significantly more effective at prolonging conversations.
Our system is the first open-source email conversation
system specialized for automatic scam-baiting. The system
aims at utilizing general NLP models for the purpose of
consuming scammers’ resources.
2.2. Text generation for email conversations
The recent successes in natural language processing
(NLP) have given rise to the prosperity of automatic
dialogue systems [7]. The most prominent architectures
include the Transformer [58] and its variants BERT [12]
and the GPT family [44]. The emergence of transformers
has enabled the pretraining-finetuning approach in NLP,
which was not possible in the era of RNN [48] and
LSTM [23]. We briefly introduce these models in this
section.