
uate models that exploit a few known biases and
find that examples from annotators who use cogni-
tive heuristics are more easily solvable by biased
models. We also examine what impact heuristics
have on trained models. Previous work (Geva et al.,
2019) shows that models generalize poorly when
datasets are split randomly by annotators, likely
due to the existence of artifacts. We replicate this
result and find that models generalize even worse
when trained on examples from heuristic-seeking
annotators.
To understand which parts of the annotation
pipeline contribute to heuristic-seeking behavior in
annotators, we also tease apart the effect of compo-
nents inherent to the task (e.g., passage difficulty)
as opposed to the annotators themselves (e.g., an-
notator fatigue) (§6). Unfortunately, we don’t dis-
cover simple predictors (i.e., passage difficulty) of
when annotators are likely to use heuristics.
A qualitative analysis of the collected data re-
veals that heuristic-seeking annotators are more
likely to create examples that are not valid, and
require simpler word-matching on explicitly stated
information (§7). Crucially, this suggests that mea-
surements of heuristic usage, such as those exam-
ined in this paper, can provide a general method
to find unreliable examples in crowdsourced data,
and direct our search for discovering artifacts in the
data. Because we implicate heuristic use in terms
of robustness and data quality, we suggest future
dataset creators track similar features and evaluate
model sensitivity to annotator heuristic use.1
2 Background and Related Work
Cognitive Heuristics.
The study of heuristics in
human judgment, decision making, and reasoning
is a popular and influential topic of research (Si-
mon,1956;Tversky and Kahneman,1974). Heuris-
tics can be defined as mental shortcuts, that we use
in everyday tasks for fast decision-making. For
example, Tversky and Kahneman (1974) asked par-
ticipants whether more English words begin with
the letter Kor contain Kas the
3rd
letter, and more
than 70% participants chose the former because
words that begin with Kare easier to recall, al-
though that is incorrect. This is an example of
the availability heuristic. Systematic use of such
heuristics can lead to cognitive biases, which are
irrational patterns in our thinking.
1
Our code and collected data is available at
https://github.com/chaitanyamalaviya/annotator-heuristics.
At first glance, it may seem that heuristics are
always suboptimal, but previous work has argued
that heuristics can lead to accurate inferences under
uncertainty, compared to optimization (Gigerenzer
and Gaissmaier,2011). We hypothesize that heuris-
tics can play a considerable role in determining
data quality and their impact depends on the exact
nature of the heuristic. Previous work has shown
that crowdworkers are susceptible to cognitive bi-
ases in a relevance judgement task (Eickhoff,2018),
and has provided a checklist to combat these biases
(Draws et al.,2021). In contrast, our work focuses
on how potential use of such heuristics can be mea-
sured in a writing task, and provides evidence that
heuristic use is linked to model brittleness.
Features of annotator behavior have previously
been useful in estimating annotator task accuracies
(Rzeszotarski and Kittur,2011;Goyal et al.,2018).
Annotator identities have also been found to influ-
ence their annotations (Hube et al.,2019;Sap et al.,
2022). Our work builds on these results and esti-
mates heuristic use with features to capture implicit
clues about data quality.
Mitigating and discovering biases.
The pres-
ence of artifacts or biases in datasets is well-
documented in NLP, in tasks such as natural lan-
guage inference, question answering and argu-
ment comprehension (Gururangan et al.,2018;
McCoy et al.,2019;Niven and Kao,2019,in-
ter alia). These artifacts allow models to solve
NLP problems using unreliable shortcuts (Geirhos
et al.,2020). Several researchers have proposed
approaches to achieve robustness against known
biases. We refer the reader to Wang et al. (2022)
for a comprehensive review of these methods. Tar-
geting biases that are unknown continues to be a
challenge, and our work can help find examples
which are likely to contain artifacts, by identifying
heuristic-seeking annotators.
Prior work has proposed methods to discover
shortcuts using explanations of model predictions
(Lertvittayakumjorn and Toni,2021), including
sample-based explanations (Han et al.,2020) and
input feature attributions (Bastings et al.,2021;
Pezeshkpour et al.,2022). Other techniques that
can be helpful in diagnosing model biases include
building a checklist of test cases (Ribeiro et al.,
2020;Ribeiro and Lundberg,2022), constructing
contrastive (Gardner et al.,2020) or counterfactual
(Wu et al.,2021) examples and statistical tests (Gu-
rurangan et al.,2018;Gardner et al.,2021). Our