the prior specific methods. ASDOT draws inspira-
tion from how humans describe data, namely by
first disambiguating and understanding the data
content, and then fusing and organizing the infor-
mation together into text paragraphs. As a result,
given input data (e.g., a table or graph), ASDOT
consists of two intuitive steps, i.e., data disam-
biguation and sentence fusion. Importantly, each
of the two steps is amenable to be solved with the
appropriate off-the-shelf pretrained LMs with op-
tional finetuning, enabling the unique flexibility
of ASDOT in the presence of any-shot training ex-
amples. More specifically, in data disambiguation
aiming to understand each data entry (e.g., triple
<Fearless, time, 2008>
), we use the prompted
GPT-3 model (Radford et al.,2019), which has en-
coded rich commonsense and world knowledge, to
convert the triple into a short sentence (
Fearless
was released in 2008
) with greatly reduced
ambiguity. The subsequent sentence fusion stage
then uses another LM, such as T5 (Raffel et al.,
2020), to combine all the resulting sentences into
a coherent paragraph as the final description. The
sentence fusion as a sub-task allows us to incor-
porate any available in-/out-of-domain training ex-
amples as well as existing large weakly supervised
corpus (Kasner and Dusek,2022) to finetune the
LM and boost the performance.
We evaluate the proposed approach in a wide
range of practical any-shot scenarios, including
(1) the zero-/few-/full-shot setting where we have
access to a varying number of training examples,
(2) the unseen-predicates setting where we describe
the data of new predicates that are never seen in
the training examples, and (3) the out-of-domain
setting where we are presented only with examples
from other domains. Extensive experiments show
that our approach consistently achieves significant
gains over the diverse previous methods specifically
designed for each of the different scenarios.
2 Related Work
Data-to-text (D2T) generation is a long-standing
problem in natural language processing with broad
applications in practice. Early research on this task
focused on rule-based and pipeline approaches (Ku-
kich,1983b;Reiter and Dale,1997), decomposing
the task into text planning, sentence planning, and
linguistic realisation. Recent work has developed
various neural approaches. Lebret et al. (2016b)
used a neural encoder-decoder for the task, fol-
lowed by attention (Bahdanau et al.,2015), content
selection (Puduppully et al.,2019a), entity mod-
eling (Puduppully et al.,2019b), and style imita-
tion (Lin et al.,2020) for further improved per-
formance. Recent studies have also incorporated
pretrained LMs (Kale and Rastogi,2020b;Ribeiro
et al.,2021;Clive et al.,2021). Although previous
fully-supervised methods have achieved remark-
able performances, most of them require a large
amount of in-domain training examples, leading
to limited applicability to the common low-data
scenarios in practice.
Recent interests are aroused in zero-/few-shot
data-to-text generation problems. Chen et al.
(2020b) first formulated the few-shot setting and
incorporated a pretrained model with a pointer gen-
erator as a solution. Chen et al. (2020a) developed a
knowledge-grounded pretrained LM for both zero-
and few-shot data-to-text generation. Gong et al.
(2020) and Chen et al. (2020b) proposed to solve
the few-shot task with content matching and pro-
totype memory, respectively. There are also stud-
ies on combining templates and pretrained LM for
zero-/few-shot generation. For example, Kale and
Rastogi (2020a) trained a neural model to rewrite
templates for few-shot task-oriented dialogue. Hei-
dari et al. (2021) applied the idea of template rewrit-
ing to build a practical few-shot data-to-text system.
Most of the previous methods have each focused on
a specific setting (e.g., either zero- or few-shot). In
comparison, our work studies a wide spectrum of
any-shot scenarios with a varying number of train-
ing examples from current or different domains. Of
particular relevance to our work is the approach by
Kasner and Dusek (2022), which performs zero-
shot data-to-text generation by rephrasing given
templates. However, the approach relies on human-
written templates for data disambiguation and thus
has limited applicability to wide domains. Besides,
the approach involves several components (order-
ing, aggregation, compression) to fuse sentences,
which restricts the use of any-shot examples for
improvement. The approach thus studies only in
zero-shot settings, while our work makes a compre-
hensive study on the diverse any-shot problems.
3 Any-Shot Data-to-Text Generation
We propose ASDOT for any-shot data-to-text gen-
eration. §3.1 describes the any-shot problems. We
then provide an overview of our method (§3.2) and