
Mutual Information Alleviates Hallucinations in Abstractive
Summarization
Liam van der Poel Ryan Cotterell Clara Meister
lvander@ethz.ch {ryan.cotterell,clara.meister}@inf.ethz.ch
Abstract
Despite significant progress in the quality of
language generated from abstractive summa-
rization models, these models still exhibit the
tendency to hallucinate, i.e., output content not
supported by the source document. A num-
ber of works have tried to fix—or at least un-
cover the source of—the problem with limited
success. In this paper, we identify a simple
criterion under which models are significantly
more likely to assign more probability to hallu-
cinated content during generation: high model
uncertainty. This finding offers a potential ex-
planation for hallucinations: models default to
favoring text with high marginal probability,
i.e., high-frequency occurrences in the train-
ing set, when uncertain about a continuation.
It also motivates possible routes for real-time
intervention during decoding to prevent such
hallucinations. We propose a decoding strat-
egy that switches to optimizing for pointwise
mutual information of the source and target
token—rather than purely the probability of
the target token—when the model exhibits un-
certainty. Experiments on the XSUM dataset
show that our method decreases the probability
of hallucinated tokens while maintaining the
ROUGE and BERTS scores of top-performing
decoding strategies.
https://github.com/VanderpoelLiam/
CPMI
1 Introduction
Abstractive summarization, the task of condensing
long documents into short summaries, has a num-
ber of applications, such as providing overviews of
news articles or highlighting main points in tech-
nical documents. Abstractive summarization is
usually performed using probabilistic text gener-
ators (Goyal and Durrett,2020;Mao et al.,2020;
Kryscinski et al.,2020), which have shown a strong
ability to produce fluent, human-like text (Baevski
and Auli,2019;Radford et al.,2019;Brown et al.,
2020). However, these models have been observed
to
hallucinate
facts, i.e., add information to the out-
put that was not present in the original text. This
behavior is problematic, as presenting users with
unsubstantiated content can lead to undesirable ef-
fects, such as the spread of misinformation (Bender
et al.,2021;Abid et al.,2021;Liang et al.,2021).
Some works have attributed this phenomenon to
the specific training corpora for these models, in
which ground-truth summaries often contain out-
side information that may not have been directly
deducible from the original text (Maynez et al.,
2020;Zhou et al.,2021). Others have pointed to
model architectures or training strategies (Voita
et al.,2021;Wang and Sennrich,2020;Kang and
Hashimoto,2020). While these works have given
us an improved understanding of the cause of hal-
lucinations, there still does not exist an efficient
and robust set of techniques for identifying and
preventing them during the generation process.
This work aims to first provide a simple criterion
indicating when a model is more likely to assign
higher probability to content not necessarily de-
rived from the source document. Specifically, we
link the start of a hallucination during generation to
high model uncertainty about the next token, which
we quantify by conditional entropy. We hypothe-
size that hallucinations may be due to a tendency of
models to default to placing probability mass on to-
kens that appeared frequently in the training corpus,
a behavior by language models previously observed
in several natural language processing (NLP) tasks
(Kobayashi et al.,2020;Wei et al.,2021). As a con-
sequence, generations with hallucinations would
still be viable candidates, as standard decoding
strategies for summarization optimize purely for
the probability of the generation. We propose an al-
ternative decoding strategy to combat this behavior:
When a model exhibits high uncertainty, we change
our decoding objective to pointwise mutual infor-
mation between the source document and target
token (PMI; Li et al.,2016;Takayama and Arase,
2019), encouraging the model to prioritize tokens
relevant to the source document. While changing
arXiv:2210.13210v2 [cs.CL] 29 Oct 2022