Multi-Type Conversational Question-Answer Generation
with Closed-ended and Unanswerable Questions
Seonjeong Hwang1, Yunsu Kim1,2, Gary Geunbae Lee1,2,
1Graduate School of Artificial Intelligence, POSTECH, Pohang, South Korea
2Computer Science and Engineering, POSTECH, Pohang, South Korea
{seonjeongh, yunsu.kim, gblee}@postech.ac.kr
Abstract
Conversational question answering (CQA) fa-
cilitates an incremental and interactive under-
standing of a given context, but building a
CQA system is difficult for many domains due
to the problem of data scarcity. In this paper,
we introduce a novel method to synthesize data
for CQA with various question types, includ-
ing open-ended, closed-ended, and unanswer-
able questions. We design a different genera-
tion flow for each question type and effectively
combine them in a single, shared framework.
Moreover, we devise a hierarchical answerabil-
ity classification (hierarchical AC) module that
improves quality of the synthetic data while
acquiring unanswerable questions. Manual in-
spections show that synthetic data generated
with our framework have characteristics very
similar to those of human-generated conver-
sations. Across four domains, CQA systems
trained on our synthetic data indeed show good
performance close to the systems trained on
human-annotated data.
1 Introduction
Conversational question answering (CQA) aims to
answer a question based on a given passage and
previous conversation. Unlike single-turn ques-
tion answering (QA) (Rajpurkar et al.,2016,2018;
Kwiatkowski et al.,2019), CQA encourages ques-
tioners to incrementally make follow-up questions,
which is suitable for services that require active in-
teraction between humans and systems. However,
manually creating large amounts of conversations
is very costly, which is a barrier to its utilization in
various domains.
To alleviate this issue, a few methods for con-
versational question generation have been studied
(Gao et al.,2019;Pan et al.,2019;Nakanishi et al.,
2019;Shen et al.,2021;Gu et al.,2021). Fur-
thermore, we have proposed approaches for auto-
matically synthesizing multi-turn conversational
question-answer (Q–A) pairs in order to build train-
ing data for CQA in our previous studies (Hwang
and Lee,2021,2022). However, our previous
frameworks generate only open-ended questions
that cannot be answered succinctly. In real-world
situations, concise answers, such as yes,no, and
unknown, are essential for fast interaction and sim-
plified conversations.
In this paper, we introduce MultiCQAG, a frame-
work that can generate multiple types of CQA data.
To enable this, we insert a generation flow for
closed-ended Q–A pairs to our previous framework
(Hwang and Lee,2022). We also design a hier-
archical answerability classification (hierarchical
AC) module that collects yet another type of data
— unanswerable questions — while improving data
quality by removing invalid Q–A pairs.
In experiments, CQA systems trained on our
synthetic datasets achieve an average F1 score of
77.2% for four new domains, showing a differ-
ence of only 5.4% from those trained on human-
annotated data. Moreover, we show by manual
evaluation that our synthetic data have a data distri-
bution similar to that of human-annotated data.
The contributions of this work can be summa-
rized as follows:
•
We propose MultiCQAG, which synthesizes a
CQA data consisting of various types of ques-
tions, including open-ended, closed-ended,
and unanswerable questions.
•
We design a hierarchical AC algorithm that fil-
ters out invalid Q–A pairs and acquires unan-
swerable questions.
2 Background
In our previous study, we proposed a conversa-
tional question-answer generation (CQAG) frame-
work that automatically synthesized data for CQA
given passages and that consisted of two modules:
contextual answer extraction (CAE) and conversa-
tional question generation (CQG) (Hwang and Lee,
arXiv:2210.12979v1 [cs.CL] 24 Oct 2022