in the closed-book setting, this paper proposes a
new QG model with two unique characteristics: (i)
a contrastive learning loss designed to better under-
stand the semantics of the answers and the seman-
tic relationship between answers and ground-truth
questions at a contextual-level; and (ii) an answer
reconstruction loss designed to measure the an-
swerability of the generated question. Contrastive
learning has shown promising results in many NLP
tasks, e.g., (Giorgi et al.,2021;Gao et al.,2021;
Yang et al.,2021) and aligns positive pairs bet-
ter with available supervised signals (Gao et al.,
2021); here we show how to learn question rep-
resentations by distinguishing features of correct
question-answer pairs from features of incorrectly
linked question-answer pairs. Further, to ensure the
generated questions are of good quality and can be
answered by the answer that is used for question
generation, we frame the model as a generation-
reconstruction process (Cao et al.,2019;Zhu et al.,
2020), by predicting the original answers given
the generated questions by a pre-trained seq2seq
model. In addition, we introduce a new closed-
book dataset with long-form abstractive answers –
WikiCQA – to complement existing datasets like
GooAQ (Khashabi et al.,2021) and ELI5 (Fan
et al.,2019) and show how to leverage our model
to generate synthetic data to improve closed-book
question-answering tasks.
Through experiments, we find that the proposed
QG model shows improvement through both auto-
matic and human evaluation metrics on WikiCQA
and two public datasets. Compared to the base-
line, the proposed QG framework shows an im-
provement of up to 2.0%, 2.7%, and 1.8% on
the ROUGE-L score on WikiCQA, GooAQ-S, and
ELI5, respectively, and 1.3% and 2.6% in terms of
relevance and correctness. Furthermore, we lever-
age the QG framework to generate synthetic QA
data from WikiHow summary data and pre-train
a closed-book QA model on it in both an unsu-
pervised and semi-supervised setting. The perfor-
mance is evaluated on both seen (WikiCQA) and
unseen (GooAQ-S, ELI5) datasets. We find consis-
tent improvements across these datasets, indicating
the QG model’s effectiveness in enhancing closed-
book question-answering tasks.
In conclusion, our contributions can be summa-
rized as follows:
•
We propose a contrastive QG model, which to
our knowledge is the first work to explore con-
trastive learning for QG under a closed-book
setting.
•
The proposed model outperforms baselines on
three datasets. The human evaluation also indi-
cates that the questions generated by our model
are more informative compared to other base-
lines.
•
We leverage the QG model as a data augmenta-
tion strategy to generate large-scale QA pairs.
Consistent improvements shown on both seen
datasets and unseen datasets indicate the QG
model’s effectiveness in enhancing closed-book
question-answering tasks.
2 Related Work
Many previous works on QG are under the open-
book setting, which takes factoid short answers (Ra-
jpurkar et al.,2016) or human-generated short
answers (Koˇ
ciský et al.,2018) with the corre-
sponding passages to generate questions (Zhang
et al.,2021). Early approaches for question genera-
tion rely on rule-based methods (Labutov et al.,
2015;Khullar et al.,2018). To bypass hand-
crafted rules and sophisticated pipelines in QG,
Du et al. (2017) introduce a vanilla RNN-based
sequence-to-sequence approach with an attention
mechanism. The recently proposed pre-trained
transformer-based frameworks (Lewis et al.,2020;
Raffel et al.,2020) also improve the performance
of QG. In addition, Sultan et al. (2020) shows that
the lexical and factual diversity of QG provides
better QA training. However, their success can not
directly adapt to the closed-book setting, where
the model is supposed to generate questions solely
relying on answers. In this work, we explore the
widely applicable closed-book QG setting, which
is still under-explored.
Contrastive Learning
aims to pull semantically
similar neighbors close and push non-neighbors
apart. It has achieved great success under both
supervised and unsupervised settings. In pioneer
works, the contrastive loss function (Hadsell et al.,
2006;Chopra et al.,2005) has been proposed as a
training objective in deep metric learning consid-
ering both similar and dissimilar pairs. Recently,
Chen et al. (2020) proposes the SimCLR frame-
work to learn useful visual representations. View-
ing contrastive learning as dictionary look-up, He
et al. (2020) present Momentum Contrast (MoCo)
to build dynamic dictionaries for contrastive learn-
ing. Some works apply contrastive learning into