detection, aiming to build a system that can work in
the real world without any specific attention to the
text domains or topic forms. More importantly, no
task-specific supervision is needed. To achieve this,
we propose to combine two types of supervision:
indirect supervision and weak supervision. The in-
direct supervision comes from textual entailment—
we treat the stance detection problem as a textual
entailment task since the attitude toward a topic
should be inferred from the input text. Therefore,
the existing entailment datasets, such as MNLI
(Williams et al.,2018), can contribute supervision
to the zero-shot setting. To collect supervision
that is more specific to the
OpenStance
task,
we design two MASK choices (
MASK-topic
and
MASK-text
) to prompt GPT-3 (Brown et al.,
2020) to generate weakly supervised data. Given an
input text and a stance label (
support
,
oppose
,
or
neutral
),
MASK-topic
predicts what topic
is appropriate based on the content; given a topic
and a label,
MASK-text
seeks the text that most
likely holds this stance. The collection of weakly
supervised data only needs the unlabeled texts and
the set of topics that users want to include. The
joint power of indirect supervision and weak super-
vision will be evaluated on VAST, SemT6 and Per-
spectrum (Chen et al.,2019), three popular datasets
that cover distinct domains, different sizes and di-
verse textual forms of topics. Experimental results
show that although no task-specific supervision is
used, our system can get robust performance on all
three datasets, even outperforming the task-specific
supervised models (72.6 vs. 69.3 by mean F1 over
the three datasets).
Our contributions are threefold: (i) we define
OpenStance
, an open-domain zero-shot stance
detection task, that fulfills real-world requirements
while having never been studied before; (ii) we de-
sign a novel masking mechanism to let GPT-3 gen-
erate weakly supervised data for
OpenStance
.
This mechanism can inspire other NLP tasks that
detect relations between two pieces of texts; (iii)
our approach, integrating indirect supervision and
weak supervision, demonstrates outstanding gen-
eralization among three datasets that cover a wide
range of text domains, topic sizes and topic forms.
2 Related Work
Stance detection.
Stance detection, as a recent
member of the NLP family, was mainly driven by
newly created datasets. In the past studies, datasets
have been constructed from diverse domains like
online debate forums (Walker et al.,2012b;Hasan
and Ng,2014;Abbott et al.,2016), news comments
(Krejzl et al.,2017;Lozhnikov et al.,2018), Twitter
(Mohammad et al.,2016;Küçük,2017;Tsakalidis
et al.,2018)), etc.
Zero-shot stance detection.
Recently, re-
searchers started to work on zero-shot stance
detection in order to build a system that can handle
unseen topics. Most work split the collected
topic-aware annotations into train and test within
the same domain. Allaway and McKeown (2020)
made use of topic similarity to connect unseen top-
ics with seen topics. Allaway et al. (2021) designed
adversarial learning to learn domain-independent
information and topic-invariant representations.
Similarly, Wang and Wang (2021) applied
adversarial learning to extract stance-related but
domain-invariant features existed among different
domains. Liu et al. (2021) utilized common sense
knowledge from ConceptNet (Speer et al.,2017) to
introduce extra knowledge of the relations between
the texts and topics. Most prior systems worked on
a single domain and were tested on a small number
of unseen topics. Li et al. (2021) tried to test on
various unseen datasets by jointly optimizing on
multiple training datasets. However, they still
assumed that part of the topics or domains has rich
annotations. In contrast, our goal is to design a
system that can handle stance detection in an open
world without requiring any domain constraints or
topic-specific annotations.
Textual entailment as indirect supervision.
Textual entailment studies if a hypothesis can be
entailed by a premise; this was proposed as a uni-
fied inference framework for a wide range of NLP
problems (Dagan et al.,2005). Recently, textual en-
tailment is widely utilized to help solve many tasks,
such as few-shot intent detection (Xia et al.,2021),
ultra-fine entity typing (Li et al.,2022), corefer-
ence resolution (Yin et al.,2020), relation extrac-
tion (Xia et al.,2021;Sainz et al.,2021), event
argument extraction (Sainz et al.,2022), etc. As
far as we know, our work is the first one that suc-
cessfully leverages the indirect supervision from
textual entailment for stance detection.
Weak supervision from GPT-3.
As the cur-
rently most popular and (arguably) well-behaved
pre-trained language model, GPT-3 (Brown et al.,
2020) has been a great success on few-shot and