lence detection. Our approach to the identification
of semantic equivalence moves beyond surface-
level perception of divergence and accounts for
more subtle differences in meaning that cannot be
detected by simply comparing the words of the sen-
tences. We leverage the Abstract Meaning Repre-
sentation (AMR; Banarescu et al.,2013) formalism
to identify these divergences at a fine-grained level.
We present an approach to semantic equivalence
detection via AMR, with analysis of data in two lan-
guage pairs (English-French and English-Spanish).
Drawing on previous work on semantic divergence
(§2), we demonstrate that sentence-level diver-
gence annotations can be coarse-grained, neglect-
ing slight differences in meaning (§3). We find that
comparing two AMR graphs is an effective way
to characterize meaning in order to uncover finer-
grained divergences (§4), and this can be achieved
even with automatic AMR parsers (§5). Finally, in
§6we evaluate our AMR-based metric on a cross-
linguistic semantic textual similarity dataset, and
show that for detecting semantic equivalence, it is
more precise than a popular existing model, multi-
lingual BERTScore (Zhang et al.,2020).
Our primary contributions include:
•
Our novel approach to the identification of se-
mantic divergence—which moves beyond per-
ceived sentence-level divergences—through
the lens of AMR
•
A simple pipeline algorithm that modifies
Smatch to quantify divergence in cross-lingual
AMR pairs, automating the detection of AMR-
level divergence
•
Studies demonstrating that our AMR-based
approach accurately captures a finer-grained
degree of semantic equivalence than both the
state-of-the-art corpus filtering method and a
semantic textual metric
We will release the code and dataset for this work
upon publication to enable the use of AMR for
semantic divergence detection.
2 Background on Semantic Divergence
Semantic divergences can arise when translating
from one language or another. These divergences
can arise due to different language structure, syn-
tactic differences in the language, or translation
choices (Dorr,1994,1990). Additional divergences
can be introduced when automatically extracting
and aligning parallel resources (Smith et al.,2010;
Zhai et al.,2018;Fung and Cheung,2004).
To address these divergences, a number of sys-
tems have been developed to automatically iden-
tify divergences in parallel texts (Carpuat et al.,
2017;Vyas et al.,2018;Briakou and Carpuat,2020,
2021;Zhai et al.,2020). The approach taken by
Briakou and Carpuat (2020) to detecting sentence-
level semantic divergences involves training mul-
tilingual BERT (Devlin et al.,2018) to rank sen-
tences diverging to various degrees. They intro-
duced and evaluated on a novel dataset called Ra-
tional English-French Semantic Divergences (RE-
FreSD). REFreSD consists of English-French paral-
lel sentences, with crowdsourced annotations clas-
sifying the sentences as having no meaning diver-
gence, having some meaning divergence, or being
unrelated.
Recent work has investigated the differences in
cross-lingual (English-Spanish) AMR pairs within
the framework of translation divergences (Wein
and Schneider,2021). Specifically, this work de-
veloped an annotation schema to classify the types
and causes of differences between cross-lingual
AMR pairs. We use this dataset to test the per-
formance of our system on English-Spanish gold
AMR pairs. (For English-French, we produce our
own gold judgments of AMR divergence to test our
algorithm.) Additional prior work has explored the
role of structural divergences in cross-lingual AMR
parsing (Blloshmi et al.,2020;Damonte,2019).
3 AMR for Identification of Semantic
Equivalence
Semantic representations are designed to capture
and formalize the meaning of a sentence. In partic-
ular, the Abstract Meaning Representation (AMR)
framework aims to formalize sentence meaning as a
graph in a way that is conducive to broad-coverage
manual annotation (Banarescu et al.,2013,2019).
These semantic graphs are rooted and labeled, such
that each node of the graph corresponds to a se-
mantic unit. AMR does not capture nominal or
verbal morphology or many function words, ab-
stracting away from the syntactic features of the
sentence. Attributes are labeled on edges between
nodes (concepts), and these attributes can be either
core roles / arguments, marked with
:ARG0
,
:ARG1
,
etc., or non-core roles such as
:time
,
:domain
, and
:manner
. The root of the AMR is usually the main
predicate of a sentence.
We leverage the semantic information captured
by AMR to recognize semantic equivalence or di-