Transferring Knowledge via Neighborhood-Aware Optimal Transport
for Low-Resource Hate Speech Detection
Tulika Bose Irina Illina Dominique Fohr
Universite de Lorraine, CNRS, Inria, LORIA, F-54000 Nancy, France
{tulika.bose, illina, dominique.fohr}@loria.fr
Abstract
Warning: this paper contains content that
may be offensive and distressing.
The concerning rise of hateful content on on-
line platforms has increased the attention to-
wards automatic hate speech detection, com-
monly formulated as a supervised classifi-
cation task. State-of-the-art deep learning-
based approaches usually require a substantial
amount of labeled resources for training. How-
ever, annotating hate speech resources is ex-
pensive, time-consuming, and often harmful to
the annotators. This creates a pressing need to
transfer knowledge from the existing labeled
resources to low-resource hate speech corpora
with the goal of improving system perfor-
mance. For this, neighborhood-based frame-
works have been shown to be effective. How-
ever, they have limited flexibility. In our pa-
per, we propose a novel training strategy that
allows flexible modeling of the relative prox-
imity of neighbors retrieved from a resource-
rich corpus to learn the amount of transfer. In
particular, we incorporate neighborhood infor-
mation with Optimal Transport, which permits
exploiting the geometry of the data embedding
space. By aligning the joint embedding and
label distributions of neighbors, we demon-
strate substantial improvements over strong
baselines, in low-resource scenarios, on differ-
ent publicly available hate speech corpora.
1 Introduction
With the alarming spread of Hate Speech (HS) in
social media, Natural language Processing tech-
niques have been used to develop automatic HS
detection systems, typically to aid manual con-
tent moderation. Although deep learning-based
approaches (Mozafari et al.,2019;Badjatiya et al.,
2017) have become state-of-the-art in this task,
their performance depends on the size of the la-
beled resources available for training (Lee et al.,
2018;Alwosheel et al.,2018).
Annotating a large corpus for HS is considerably
time-consuming, expensive, and harmful to human
annotators (Schmidt and Wiegand,2017;Malmasi
and Zampieri,2018;Poletto et al.,2019;Sarwar
et al.,2022). Moreover, models trained on existing
labeled HS corpora have shown poor generaliza-
tion when evaluated on new HS content (Yin and
Zubiaga,2021;Arango et al.,2019;Swamy et al.,
2019;Karan and Šnajder,2018). This is due to the
differences across these corpora, such as sampling
strategies (Wiegand et al.,2019), varied topics of
discussion (Florio et al.,2020;Saha and Sindhwani,
2012), varied vocabularies, and different victims
of hate. Thus, to address these challenges, here we
aim to devise a strategy that can effectively trans-
fer knowledge from a resource-rich source corpus
with a higher amount of annotated content to a low-
resource target corpus with fewer labeled instances.
One popular way to address this is transfer learn-
ing. For instance, Mozafari et al. (2019) fine-tune
a large-scale pre-trained language model, BERT
(Devlin et al.,2019), on the limited training exam-
ples in HS corpora. Further, a sequential trans-
fer, following Garg et al. (2020), can be per-
formed where a pre-trained model is first fine-
tuned on a resource-rich source corpus and sub-
sequently fine-tuned on the low-resource target cor-
pus. Since this may risk forgetting knowledge from
the source, the source and target corpora can be
mixed for training (Shnarch et al.,2018). Besides,
to learn target-specific patterns without forgetting
the source knowledge, Meftah et al. (2021) aug-
ment pre-trained neurons from the source model
with randomly initialized units for transferring
knowledge to low-resource domains.
Recently, Sarwar et al. (2022) argue that tradi-
tional transfer learning strategies are not systematic.
Therefore, they model the relationship between a
source and a target corpus with a neighborhood
framework and show its effectiveness in transfer
learning for content flagging. They model the in-
arXiv:2210.09340v1 [cs.CL] 17 Oct 2022