
information (Janowicz et al.,2018) and leads to
higher barriers for female personalities to receive
a biographic entry (Beytía et al.,2022). In an ex-
periment, Demartini (2019) asked crowd contribu-
tors to provide a factual answer to the (politically
charged) question of whether or not Catalonia is a
part of Spain. The diverging responses indicated
that participants’ beliefs of what counts as true dif-
fered largely. This is an example of bias that is
beyond a subliminal psychological level. In this
case, structural aspects like consumed media and
social discourse play an important role. To counter
this problem, Demartini (2019) suggests actively
asking contributors for evidence supporting their
statements, as well as keeping track of their de-
mographic backgrounds. This makes underlying
motivations and possible sources for bias traceable.
3.2 Ontologies: Manual Creation of Rules
Ontologies determine rules regarding allowed types
of entities and relations or their usage. They are of-
ten hand-made and a source of bias (Janowicz et al.,
2018) due to the influence of opinions, motivations,
and personal choices (Keet,2021): Factors like sci-
entific opinions (e.g., historical ideas about race),
socio-culture (e.g., how many people a person can
be married to), or political and religious views (e.g.,
classifying a person of type X as a terrorist or a
protestor) can proximately lead to an encoding of
social bias. Also structural constraints like the on-
tologies’ granularity levels can induce bias (Keet,
2021). Furthermore, issues can arise from the types
of information used to characterize a person entity.
Whether one attributes the person with their skin
color or not could theoretically determine the emer-
gence of racist bias in a downstream application
(Paparidis and Kotis,2021). Geller and Kollapally
(2021) give a practical example for detection and
alleviation of ontology bias in a real-world scenario.
The authors discovered that ontological gaps in the
medical context lead to an under-reporting of race-
specific incidents. They were able to suggest coun-
termeasures based on a structured analysis of real
incidents and external terminological resources.
3.3 Extraction: Automated Extraction of
Information
Natural language processing (NLP) methods can
be used to recognize and extract entities (named
entity recognition; NER) and their relations (rela-
tion extraction; RE), which are then represented
as
[
head entity, relation, tail entity
]
tuples (or as
[subject, predicate, object], respectively).
Mehrabi et al. (2020) showed that the NER sys-
tem CoreNLP (Manning et al.,2014) exhibits bi-
nary gender bias. They used a number of tem-
plate sentences, like "<Name> is going to school"
or "<Name> is a person" using male and female
names
5
from 139 years of census data. The model
returned more erroneous tags for female names.
Similarly, Mishra et al. (2020) created synthetic
sentences from adjusted Winogender (Rudinger
et al.,2018) templates with names associated with
different ethnicities and genders. A range of dif-
ferent NER systems were evaluated (bidirectional
LSTMs with Conditional Random Field (BiLSTM
CRF) (Huang et al.,2015) on GloVe (Pennington
et al.,2014), ConceptNet (Speer et al.,2017) and
ELMo (Peters et al.,2017) embeddings, CoreNLP,
and spaCy
6
NER models). Across models, non-
white names yielded on average lower performance
scores than white names. Generally, ELMo ex-
hibited the least bias. Although ConceptNet is
debiased for gender and ethnicity
7
, it was found to
produce strongly varied accuracy values.
Gaut et al. (2020) analyzed binary gender bias
in a popular open-source neural relation extraction
(NRE) model, OpenNRE (Han et al.,2019). For
this purpose, the authors created a new dataset,
named WikiGenderBias (sourced from Wikipedia
and DBpedia). All sentences describe a gendered
subject with one of four relations: spouse,hyper-
nym,birthData, or birthPlace (DBpedia mostly
uses occupation-related hypernyms). The most no-
table bias found was the spouse relation. It was
more reliably predicted for male than female en-
tities. This observation stands in contrast to the
predominance of female instances with spouse rela-
tion in WikiGenderBias. The authors experimented
with three different mitigation strategies: down-
sampling the training data to equalize the number
of male and female instances, augmenting the data
by artificially introducing new female instances,
and finally word embedding debiasing (Bolukbasi
et al.,2016). Only downsampling facilitated a re-
duction of bias that did not come at the cost of
model performance.
Nowadays, contextualized transformer-based en-
5
While most of the works presented here refer to gender as
a binary concept, this does not agree with our understanding.
We acknowledge that gender is continuous and technology
must do this reality justice.
6https://spacy.io/
7
https://blog.conceptnet.io/posts/2017/conceptnet-
numberbatch-17-04-better-less-stereotyped-word-vectors/