
questions. For example, in Figure 1, to answer
the question “Who was the first president of the
US during WW2?” the edge with Bill Clinton has
little relevance for answering the question. But,
regular RGCN would still weigh all edges equally.
We address this shortcoming through our proposed
modulation. We impose soft temporal constraints
on the messages passed during convolution, am-
plifying messages through edges close to the time
period relevant for answering the question while
diminishing messages from irrelevant edges. This
leads to better, more efficient learning as we are
not confusing our model with unnecessary infor-
mation, as evidenced by our significantly improved
performance without the need for any heavy engi-
neering. We explore two different strategies for our
convolutional edge weighting, which show comple-
mentary strengths. Our experiments establish that
TwiRGCN significantly outperforms already strong
baselines on TimeQuestions. Our contributions are:
•
We propose TwiRGCN, a simple and gen-
eral TKGQA system that computes question-
dependent edge weights to modulate RGCN mes-
sages, depending on the temporal relevance of
the edge to the question.
•
We explore two novel and intuitive schemes for
imposing soft temporal constraints on the mes-
sages passed during convolution, amplifying mes-
sages through edges close to the time relevant for
answering the question while diminishing mes-
sages from irrelevant edges. We also propose
an answer-gating mechanism based on the likeli-
hood that the answer is an entity or time.
•
Through extensive experiments on a challeng-
ing real-world dataset, we find that TwiRGCN
substantially outperforms prior art in overall ac-
curacy, and by 9–10% on the implicit and ordinal
type questions — categories that require signifi-
cant temporal reasoning.
•
We augment TimeQuestions with a TKG and re-
lease both code and data at https://github.com/adi-
sharma/TwiRGCN.
2 Related Work
Most KGQA systems have focused on answering
questions from simple (i.e., 1-hop fact-based ques-
tions) (Berant et al.,2013) to multi-hop complex
questions requiring multi-fact reasoning (Sun et al.,
2019;Saxena et al.,2020). However, only a small
fraction of these questions require any temporal
reasoning (Jia et al.,2018a). Recent efforts have
President of the
United States
position held
[1913, 1921]
position held
[1993, 2001]
position held
[1945, 1953]
Franklin D.
Roosevelt
position held
[1933, 1945]
Bill Clinton
Harry
Truman
Woodrow
Wilson
Who was the first president of US during WW2?
where
Figure 1: An illustrative example of how our temporal gating
described in Section 4.2 modulated the incoming graph con-
volution messages for one node depending on the time period
of interest for the question. The thickness of an edge here is
proportional to the value of the temporal edge weight
m(e)
tq
for that edge. In this example, the entities Franklin D. Roo-
sevelt and Harry Truman, who were presidents during WW2
[1939,1945]
get the top two highest weights, while Woodrow
Wilson, who was president during WW1
[1914,1918]
gets a
smaller edge weight. In contrast, Bill Clinton, whose time
period is unrelated to the question, gets a much lower edge
weight. Thus, contributing very little to the convolution update
of the ’President of the US’ node.
tried to overcome this gap by proposing models
as well as datasets to explicitly focus on temporal
reasoning. We review these below.
Temporal KGQA methods: One line of work
uses temporal constraints along with hand-crafted
rules to find the answer (Bao et al.,2016;Luo
et al.,2018;Jia et al.,2018b). A recent class of
models has leveraged advances in TKG embed-
ding methods for answering questions on Tempo-
ral KGs. CronKGQA (Saxena et al.,2021) does
this by posing a question as a TKG completion
problem and finds the answer using the TComplex
(Lacroix et al.,2020) score function and BERT (De-
vlin et al.,2018) question embedding to complete
the fact. TempoQR (Mavromatis et al.,2021) uses
additional temporal supervision to enrich TKG em-
beddings, followed by a transformer-based decoder
(Vaswani et al.,2017). TSQA (Shang et al.,2022)
on the other hand estimate the time in the question
and uses it to enrich TKG embeddings for finding
the answer. SubGTR (Chen et al.,2022) infers
question-relevant temporal constraints using TKG
embeddings and applies them as filters to score en-
tities in the question subgraph. Although we, too,
use pre-trained TKG embeddings to initialize our
generalized RGCN, we use the GNN framework to
take advantage of the structural information in the
KG in ways that they do not. Recent work (Teru
et al.,2020) shows that GNN-based models can
encode any logical rule corresponding to a path in