
CIKM ’22, October 17–21, 2022, Atlanta, GA, USA Endri Kacupaj, Kuldeep Singh, Maria Maleshkova, & Jens Lehmann
entities for top positions (i.e., rank KG paths). The two approaches
follow either a parse-then execute paradigm or a retrieval-and-
rank paradigm. For ConvQA over KGs, there has been signicant
progress on semantic parsing-based approaches [
17
,
33
,
38
]. How-
ever, collecting training data for semantic parsing approaches is
challenging, and time-consuming [
23
] since each question must
be associated with a gold logical form. While for the information-
retrieval/ranking-based approaches, only the correct answers (e.g.,
entities/KG paths) are required for each question.
State of the Art Limitations.
Existing ranking-based ConvQA
techniques formulate the task as a path ranking problem and pro-
pose either a heuristic approach [
10
] or a reinforcement learning
model [
21
] to learn from question reformulations (by solely rely-
ing on the user). However, these techniques have their inherent
limitations. For instance, the rule-based approaches generally re-
quire tedious manual labor to generate rules, are error-prone, and
are prone to bias [
8
]. Furthermore, existing scientic literature
points to several limitations in solely relying on users for query
reformulation, and focused on automatic query suggestion/related
query recommendations techniques [6, 7]. Firstly, entirely relying
on users creates a heavy cognitive load on them [
31
]. Secondly,
reformulation query issued within a time interval that is short
enough (such as ve minutes) implies dissatisfaction [
14
]. Lastly, a
recent study points out that critical issues for conversational sys-
tems concerning reformulations have not been widely analyzed
in the literature, especially the reformulations’ implicit ambiguity
[
22
]. Hence, for advancing IR-based ConvQA methods, there is a
desirable ask for approaches that are independent of availability
of gold-logical form, heuristic rules, or its sole dependency on the
users for query reformulation.
1.1 Proposed Approach, Motivation, and
Contributions
This paper addresses a relatively unexplored research area: the
ranking-based ConvQA task for answering conversational ques-
tions posed against a knowledge graph. In this context, we propose
PRALINE (
P
ath
R
anking for convers
A
tiona
L
quest
I
on a
N
sw
E
ring),
a novel contrastive representation learning approach to rank KG
paths for retrieving the correct answers eectively. Furthermore,
we enrich the learning process by incorporating the available con-
versational context, i.e., (1) the entire dialog history with (2) uent
responses and (3) domain information (c.f. Figure 1). Our rationale
for critical choices are following:
Why Contrastive Learning for ConvQA?
Contrastive learning [
9
,
34
] aims at learning representations of
data by contrasting similar and dissimilar samples. For our task, the
data contains conversations and respective answers (label) for each
conversational question. However, there is no discrete information
on how the answers were extracted. Therefore fully supervised
learning approaches such as semantic parsing [
33
,
38
] cannot be
directly applied without annotations. Consequently, we can design
an approach with contrastive learning by only augmenting the data
with KG paths that lead to the correct answers. These KG paths have
as starting points the context entities mentioned in the conversation
and landing points the answers. Furthermore, extracted KG paths
leading to correct answers are marked as positive, while others
are negative. In this way, contrastive learning is ideal for our task
since it allows us to rank KG paths that are considered positive and
answer conversational questions.
Why Conversational Context to Enrich Learning Process?
Conversational context plays a vital role in human understand-
ing [
13
] and question answering [
26
]. To further enhance learning,
we seek additional conversational context to improve the ranking
performance and allow contrastive learning approach to distin-
guish between positive and negative paths. Such information can
be the conversation domain and uent natural language answers
[
3
,
16
,
18
,
20
] instead of standalone KG answers (answer labels).
Identifying the domain of the conversation allows us to enrich the
representations and eciently contrast negative paths that do not
use properties of the particular domain. Moreover, uent natural
language answers will supplement the conversations with addi-
tional textual context to support the learning and ranking process.
Contributions
: We make the following key contributions in the
paper: 1) We propose PRALINE, the rst contrastive learning based
approach for ConvQA that jointly models the available conversa-
tional context (full dialog history with uent responses and domain)
and KG paths in a common space for learning joint embedding repre-
sentations to improve KG path ranking. 2) We systematically study
the impact of incorporating additional context on the performance
of PRALINE. Results on standard datasets show a considerable im-
provement over previous baselines. To facilitate reproducibility and
reuse, our framework implementation and the results are publicly
available1. The structure of the paper is as follows: Section 2 sum-
marizes the related work. Section 3 provides the concepts, notations
and tasks denitions. Section 4 presents the proposed PRALINE
framework. Section 5 describes the experiments, including the ex-
perimental setup, the results, the ablation study and error analysis.
We conclude in Section 6.
2 RELATED WORK
Considering KGQA is a widely studied research topic, we stick to
the work closely related to our proposed approach (detailed surveys
are in [12, 43]).
Single-shot KGQA.
Several KGQA works handle the task as a se-
mantic graph generation and re-ranking. Bast and Haussmann
[1]
compare a set of manually dened query templates against the natu-
ral language question and generate a set of query graph candidates
by enriching the templates with potential relations. Yih et al
. [41]
creates grounded query graph candidates using a staged heuristic
search algorithm and employs a neural ranking model to score and
nd the optimal semantic graph. Yu et al
. [42]
use a hierarchical rep-
resentation of KG relations in a neural query graph ranking model.
Authors compare the results against a local sub-sequence alignment
model with cross attention [
32
]. Maheshwari et al
. [28]
conduct an
empirical investigation of neural query graph ranking approaches
by experimenting with six dierent ranking models. The proposed
approach is a self-attention-based slot matching model that exploits
the inherent structure of query graphs.
ConvQA over KGs.
Most recent works on ConvQA [
17
,
33
,
38
]
employ the semantic parsing approach to answer conversational
questions. The rst work in this area Saha et al
. [36]
propose a
1https://github.com/endrikacupaj/PRALINE