
Jans et al.,2012;Rudinger et al.,2015;Granroth-
Wilding and Clark,2016) and later extended to
multiple protagonists (Pichotta and Mooney,2014;
Peng and Roth,2016;Pichotta and Mooney,2016;
Modi,2016;Weber et al.,2018,2020;Zhang et al.,
2020b). All of these works assume there exists a
single line of events that describes all occurrences
within a complex event. This work does not limit it-
self to generating single-chained schemas. We also
consider more complex graphs as schema outputs.
In addition, none of these works deal with zero-shot
scenarios that do not require training data.
Pre-Trained Generation Models:
Large-scale
pre-trained text generation models such as GPT-2
(Radford et al.,2019), GPT-3 (Brown et al.,2020),
BART (Lewis et al.,2020), T5 (Raffel et al.,2020),
i.a. have been used in many NLP tasks. These
models are often seen as few-shot learners (Brown
et al.,2020) and therefore used as inference meth-
ods. However, these text generation models are not
explicitly trained to perform inference, but to pro-
duce the most likely sequence of words to proceed
a certain prompt, similar to language models. In
our work, we use these large pre-trained LMs as
text generators. The generated documents on a par-
ticular topic are leveraged as a corpus for extracting
the schema of the given topic. We rely on the in-
tuition that the generated text will include salient
and stereotypical information that is expected to
be mentioned in the context of the topic (e.g., for
the topic of “planning a wedding,” we assume most
documents will include “order catering”).
3 Data Generation
The schema induction process begins with generat-
ing texts using large LMs as text generation models.
These texts are joined to form a knowledge base
for the schema, including all of the potential infor-
mation that the schema may present. One could,
of course, create this knowledge base by crawling
the web for real news articles or Wikipedia entries
related to a certain topic.
We argue, however, that in addition to the obvi-
ous advantages of not having to rely on the avail-
ability of data online and not having to crawl the
entire web for relevant documents on each topic,
the generated data from these large generative mod-
els is more efficient in reporting salient events than
random events described in the news, i.e., gener-
ated texts are more likely to mention important
information than real documents do.
Generated Text Real Text
# events / # tokens 12.52% 6.31%
# arguments / # tokens 5.45% 3.01%
Table 1: The ratio of relevant events and relevant argu-
ment roles identified in generated texts and real texts
for the scenario of IED attack.
Our analysis shows that the generated stories
contain a higher percentage of relevant tokens than
real news articles that are used for schema induc-
tion. To demonstrate this phenomenon, we com-
pare manually collected documents with those that
are automatically generated using GPT-3 for the
event of Improvised Explosive Device (IED) At-
tack (Li et al.,2021). To identify salient events and
arguments concerning IED attacks, we adopt the
DARPA KAIROS Phase 1 (v3.0) ontology
2
— a
fine-grained ontology for schema learning, with 24
entity types, 67 event types, and 85 argument roles.
We calculate the number of relevant event trig-
gers and arguments identified in the text, where a
relevant mention is one whose type appears in the
ontology. The results shown in Table 1demonstrate
that the quality of the generated texts in terms of
conciseness and appearance of important details is
higher than that of real texts. For example, the ratio
of relevant events per token is more than twice as
high in generated texts as it is in real texts. Hence
we are able to not only generate a schema for every
given topic without putting any effort in searching
the web, but the information we generate is also
better suited for our end task of depicting all of the
important aspects of a complex event.
Given a topic for which we want to create a
schema, we generate multiple texts that discuss the
topic event using the OpenAI GPT-3 API
3
with the
Davinci-instruct-beta-v3 model and we also experi-
ment with the Hugging Face GPT-2 API
4
. We use
three prompting methods to generate documents of
diverse genres as follows:
News Articles:
We begin by generating a headline
using the prompt: “Write a news headline about
topic.” The output from this prompt is then used
in the following prompt: “Write a news story titled
headline.” The output from the second prompt is
added to the pool of generated texts. The process
is repeated 30 times. See example in Fig. 2b.
2
The full ontology definition can be accessed at this link:
https://bit.ly/3mIWJoN.
3https://openai.com/blog/openai-api/.
4https://huggingface.co/gpt2