learning-to-learn or meta-learning paradigm can
further enhance the utilization of instructions by
learning about task at deeper levels. In this pa-
per, we investigate how smaller LMs could best
benefit from the natural instructions and whether
meta-learning paradigms can further improve the
zero-shot generalization ability of LMs in MTIL.
Meta-learning has been shown to be effective in
adapting knowledge with little supervision but to
the best of our knowledge has not been adapted to
MTIL in zero-shot settings.
Specifically, we explore two different meta-
learning approaches. First we propose to adapt
Model Agnostic Meta Learning (MAML) (Finn
et al.,2017) for MTIL, an optimization based
approach. Second, we explore hyper-network
(HNet) (Ha et al.,2017) based MTIL, a black-box
approach. HNet introduces an auxiliary LM which
encodes instructions to produce task specific param-
eters which are added to the main LM parameters
to generate a task specific LM at prediction time.
In addition, we evaluate a third approach which
combines the two into a HNet-MAML by training
the HNet model using MAML.
We conduct extensive experiments specifically
designed to test the generalization ability of LMs
trained with instructions under different zero shot
conditions. We use two sets of training tasks from
the NIV2 dataset: 1) all natural language tasks and
2) natural language generation tasks. We evalu-
ate the models for two sets of held out
generation
tasks
conveying different levels of zero-shot gener-
alization ability: 1)
weak generalization
set with
a random selection of generation tasks with poten-
tial overlap of categories with training tasks and 2)
strong generalization
set (or strict zero-shot con-
ditions) using summarization and title generation
tasks with no overlap in categories from the train-
ing tasks. We further investigate the task sets under
difficulty levels of easy,medium, and hard based
on their baseline ROUGE scores.
The main conclusion from our study is that under
strict zero-shot conditions, meta-learning with in-
structions significantly improves the performance.
The improvements become more significant for the
strong generalization task set and when the task
difficulty level is hard (i.e. tasks where the LM
struggles to generate correct outputs in zero-shot
setting). Moreover, meta-learning increases the
effectiveness of instructions under all conditions.
While both MAML and HNet models show im-
provements over the baselines, HNet (along with its
MAML extension) by explicitly enforcing the use
of instructions through task specific conditioning of
parameters, results in larger gains. In summary, the
main contributions of the paper are two-fold. First,
we adapt meta-learning approaches to MTIL. Sec-
ond, we study their efficacy and show significant
improvements under strict zero-shot conditions.
2 Related Work
Learning from instructions: An extension of the
basic prompt-based in-context learning is append-
ing task specific instructions with prompts. Sev-
eral recent works which include FLAN (Wei et al.,
2022), T0 (Sanh et al.,2022) and (Reif et al.,2021),
train a large LM in a multi-task setting with instruc-
tions. InstructGPT (Ouyang et al.,2022) takes
slightly different approach by training the GPT3
model (Brown et al.,2020) with human anno-
tated dataset of demonstrations of desired user in-
tents and use reinforcement learning to improve
the model to follow such instructions. Yet an-
other direction called pattern-exploiting training
(PET) (Schick and Schütze,2021a;Schick and
Schütze,2021) combines the idea of formulating
instructions as cloze questions and show that even
small LMs can be good few-shot learners and work
with language generation.
Meta-learning for language generation
: Meta
learning has been applied in several language gen-
eration settings such as (Lin and Lee,2020) to
induce persona in a chatbot, (Mi et al.,2019) for
task oriented dialog systems, (Gu et al.,2018) for
low resource machine translation, and (Chen and
Shuai,2021) for abstractive summarization in a
low-resource transfer learning but do not use in-
structions for zero-shot transfer. Our MTIL sce-
nario is closely related to MetaICL (Min et al.,
2022) which applies multi-task learning in-context
in a K-shot setting for classification tasks, but dif-
fers in that it is a k-shot in-context scenario and
does not use instructions or meta-learning optimiza-
tion. While these works are related, to the best of
our knowledge, meta-learning has not been used
to generalize to unseen generation tasks in zero
shot settings using instructions and thus the paper
provides several novel insights and approaches.
Hyper-Networks (HNet) in NLP applica-
tions
: (Karimi Mahabadi et al.,2021) use HNet
to train LMs in a multi-task setting with adapters
and (von Oswald et al.,2020) propose a contin-