
entailment among different statements while ignor-
ing the rebuttal ones, which could be crucial in
real applications. For example,
sent5
counters
sent4
as a condition of exception and we can-
not construct the correct reasoning graph without
the rebuttal relation. Second, there could exist in-
ternal logical relations inside each statement. For
example,
sent5
contains two atomic sentences
connected by a logical implication relation. Third,
real-life statements could have different degrees of
certainty. For example, “He is hungry” and “He is
likely to be hungry” are not identical but relevant
because of the certainty. However, most previous
work simply treats them completely separately in-
stead of considering their relevance and trying to
model the difference (i.e., certainty).
Motivated by previous cognitive science work
(i.e., Toulmin Model
1
(Toulmin,2003) and modal
logic theory
2
(Garson,2021)), we propose a new
explanation form, logic metagraphs, to address the
aforementioned limitations of previous work. As
demonstrated in Figure 1, the logical metagraphs
are directed acyclic graphs with meta nodes con-
nected by two types of edges, support and rebut,
representing the inferences between the statements
over a logical passage. The meta structure uncovers
the chain of reasoning from evidence to the con-
clusion, along with the challenges from the rebut-
tal sentences. Each meta node stores information
about a logically sound statement formulated as a
propositional formula in a standard modal logic S5
system (Hughes et al.,1996), a direct extension of
first-order propositional logic with two certainty
operators. The formulae have atomic sentences
as logical variables that denote events or beliefs,
which are modified by three unary operators on
their certainty (negation
¬
, necessity
2
, and possi-
bility
3
) and are joined by three binary operators
on their logical relations (implication
→
, conjunc-
tion
∧
, disjunction
∨
). As a result, the logic meta-
graphs are comprehensive with multi-hop reason-
ing paths, inference rebuttal, the internal structure
1
The Toulmin Model is a canonical theory that helps for-
mat and understand arguments. It provides a general pattern
to assign logical roles to the sentences in the argument, which
clarify the overall logical relations. Especially, the rebuttal
components challenge the derivation from existing evidence
to the conclusion by providing additional information such as
giving a counterexample or proposing an additional condition.
2
The modal logic theory extends classic first-order propo-
sitional logic with two modal operators about certainty and
several corresponding rules. This facilitates us to keep the
logical variables and relations found in the text and, at the
same time, introduce degrees of certainty to the graph.
of the statements, and reasoning strength denoted
by the degrees of certainty. We collect 1,000 log-
ical passages from the ReClor dataset (Yu et al.,
2020) and build the MetaLogic dataset.
Based on our new explanation form, we exam-
ine the current best models’ ability to understand
logical reasoning profoundly. The models need
to generate the logic metagraphs given a logical
passage. Performances are evaluated by matching
scores for the overall structure as well as the three
fine-grained components: (1) The inference steps
between meta nodes; (2) The per-statement formu-
lae with multiple logical triples; (3) The degrees of
certainty. Our evaluation results indicate that gener-
ating a comprehensive logical reasoning structure
is still challenging for existing giant models.
Our contributions are three-fold:
1.
We propose a new explanation form, the logic
metagraphs, with a comprehensive logical struc-
ture and rich logical information, and the corre-
sponding metagraph generation task.
2.
We build a high-quality dataset, MetaLogic, on
real-world logical passages.
3.
We conduct experiments on three generative
models in different frameworks and locate the
challenges for current models.
2 Related Works
Explanations
Explanation in the context of natu-
ral language understanding tasks (e.g., QA) pro-
vides interpretability about how models solve the
problem. The strategies include asking the models
to generate rationales while answering the ques-
tions (DeYoung et al.,2020;Inoue et al.,2020),
and deriving multi-hop chains of reasoning (Jham-
tani and Clark,2020;Dalvi et al.,2021). The
single-sentence rationale provides justification for
the question answering but does not uncover the
reasoning procedure. While the form of multi-hop
chains of reasoning uncovers the reasoning proce-
dure and remedies the simple justification of ra-
tionale, it still lacks critical clues about the mech-
anism within the reasoning steps. Our proposed
fine-grained explanation form extends the chain of
reasoning by unwrapping the fine-grained texture
within each reasoning step. As a result, it allows
the reasoning chains to include multiple inference
types (e.g., rebuttal) and broader reasoning types