
Author et al.: Preparation of Papers for IEEE TRANSACTIONS and JOURNALS
articles; its annotation process was mainly performed by just
one person; and some of its methodologies are subject to dis-
cussion, such as including the name of the user as a predictive
feature. Mubarak et al. [56] built a dataset of comments taken
from the Al Jazeera website,6and annotated them together
with the title of the article, but without including the entire
thread of replies.
Pavlopoulos et al. [55] analyze the impact of adding
context to the toxicity detection task. They find that, while
humans seem to leverage conversational context to detect
toxicity, the trained classification models were not able to
improve their performance significantly by adding context.
Following up, Xenos et al. [57] label each message with its
“context sensitiveness”, measured as the difference between
two groups of annotators: those who have seen the context,
and those who have not. With this, they observe that classi-
fiers improve their performance on comments which are more
sensitive to context.
Further, Sheth et al. [58] explore some opportunities for
incorporating richer information sources into the toxicity
detection task, such as the interaction history between users,
some kind of social context, and other external knowledge
bases. Wiegand et al. [59] pose some questions and chal-
lenges regarding the detection of implicit toxicity — that is,
some subtle forms of abusive language not expressed as slurs.
Summing up, BERT-based models are state-of-the-art for
this type of classification tasks; there have been various at-
tempts to include context in distinct ways and with disparate
success; there have been relatively few studies on Spanish
data; and hate speech detection has typically been addressed
as a binary task, making no distinction among the attacked
characteristics or calls-to-action. In the present work, we
assess the usefulness of adding context, we work with BERT-
based models, on Spanish data, and address both binary and
fine-grained classification tasks.
III. DEFINITION OF HATE SPEECH
We say that there is hate speech in a comment if it contains
statements of an intense and irrational nature of disapproval
and hatred against an individual or a group of people because
of its identification with a group protected by domestic
or international laws [1]. Protected treats or characteristics
include color, race, national or social origin, gender identity,
language, and sexual orientation, among others.
Hate speech can manifest itself explicitly as direct in-
sults, slurs, celebrations of crimes, incitements to take ac-
tion against an individual or group, or even more veiled
expressions such as ironic content. Following this definition,
we consider that an insult or aggression is not enough to
constitute hate speech; it is necessary to make an explicit or
implicit appeal to at least one protected characteristic.
For international law, hate speech has an extra element
that differentiates it from offensive behavior: the promotion
of violent actions against its targets. However, the NLP
6https://www.aljazeera.com/
Short name Hate speech against ...
WOMEN women
LGBTI gay, lesbian, bisexual, transgender, intersexual peo-
ple
RACISM people based on their race, skin color, language, or
national identity
CLASS people based on their socioeconomic status
POLITICS people based on their political affiliation or ideology
APPEARANCE fat people, old people, or other aspect-based features
CRIMINAL criminals and persons in conflict with law
DISABLED people with disability or mental health affections
Table 1: Protected characteristics considered in this work.
Short names are used throughout the paper to refer to these
broad groups.
community does not usually require this “call to action” when
identifying hate speech. In the present work, we will adopt
this latter view, and we will explicitly state when we also
refer to calls to action.
Several characteristics are taken into account in this work.
In addition to misogyny and racism (the most common treats
considered in previous works), we also consider: homopho-
bia and transphobia; social class hatred (sometimes known
as aporophobia); hatred due to physical appearance (e.g.,
overweight); hatred towards people with disabilities; political
hate speech; and hate speech against criminals, prisoners,
offenders and other people in conflict with the law. For this
selection, we take into account the definition of discrimina-
tion from international human rights treaties, which refers
to discrimination motivated by race, color, sex, language,
religion, political, or other opinions, national or social origin,
property, birth or other status [60]. These eight characteristics
are listed in Table 1 along with reference names that will be
used throughout the paper.
IV. CORPUS
This section describes the collection, curation, and annota-
tion process of the corpus. Our aim was to construct a dataset
of user messages commenting on specific news articles, in
a similar fashion to the reader forums present in many news
outlet websites. Figure 2 offers a schematic illustration of our
dataset, with a tweet from a news outlet about China banning
the breeding of dogs for human consumption, its respective
news article, and replies from users to the original tweet.
A. DATA COLLECTION
Our data collection process was targeted at the official Twit-
ter accounts of a selected set of Argentinian news outlets:
La Nación (@lanacion), Clarín (@clarincom), Infobae (@in-
fobae), Perfil (@perfilcom), and Crónica (@cronica). These
are the main National newspapers in the country, and attract
a vast volume of interaction on Twitter.
We considered a fixed time period of one year, starting
in March 2020. We collected the replies to each post of the
mentioned accounts using the Spritzer Twitter API, listening
to any tweet mentioning one of their usernames.
For the purpose of this work, we were only interested in the
4VOLUME 4, 2016