tain richer semantics but are seldom included by
a predefined type set. Second, PLMs are capable
of conceptual reasoning and handling the few-shot
and zero-shot dilemma (Hwang et al.,2021), since
massive knowledge has been learned during their
pre-training.
However, it is nontrivial to realize PLM-based
GET due to the following challenges: 1) Entity typ-
ing usually requires generating fine-grained types
with more semantics, which are more beneficial to
downstream tasks. However, PLMs are biased to
generate high-frequency vocabulary in the corpus
due to their primary learning principle based on
statistical associations. As a result, a typical PLM
tends to generate high-frequent but coarse-grained
types even if we carefully finetune the PLM on the
fine-grained entity typing dataset (refer to Figure 5
in Section 4). Therefore, how to guide a PLM to
generate high-quality and fine-grained types for
entities is crucial. 2) It is costly for humans to an-
notate a great number of samples with fine-grained
types. Therefore, most existing works adopt het-
erogeneous data consisting of a small portion (less
than 10%) of human-annotated data and a large
portion (more than 90%) of auto-generated low-
quality data (e.g., by distant supervision), which
greatly hurts the performance of entity typing mod-
els (Gong et al.,2021). How to train a PLM to
generate desirable types on these low-quality het-
erogeneous data is also challenging.
The difficulty of using PLMs to generate high-
quality fine-grained types based on the low-quality
heterogeneous training data motivates us to lever-
age the idea from curriculum learning (CL) (Ben-
gio et al.,2009), which better learns heterogeneous
data by ordering the training samples based on their
quality and difficulty (Kumar et al.,2019). In this
paper, we propose a CL-based strategy to train our
GET model. Specifically, we first define a fixed
curriculum instruction and partition the training
data into several subsets according to the granular-
ity and heterogeneity of samples for model train-
ing. Based on the curriculum instruction, CL can
control the order of using these training subsets
from coarse-grained and lower-quality ones to fine-
grained and higher-quality ones. However, a fixed
curriculum ignores the feedback from the training
process. Thus, we combine the predetermined cur-
riculum with self-paced learning (SPL) (Kumar
et al.,2010), which can enforce the model dynam-
ically self-adjusting to the actual learning order
according to the training loss. In this way, our CL-
based GET model can make the learning process
move towards a better global optimum upon the
heterogeneous data to generate high-quality and
fine-grained types. Our contributions in this paper
are summarized as follows:
•
To the best of our knowledge, our work is the
first to propose the paradigm of generative
entity typing (GET).
•
We propose to leverage curriculum learning to
train our GET model upon heterogeneous data,
where the curriculum can be self-adjusted
with self-paced learning.
•
Our extensive experiments on the data of dif-
ferent languages and downstream tasks justify
the superiority of our GET model.
2 Related Work
Classification-based Entity Typing
The tradi-
tional classification-based entity typing methods
can be categorized into three classes. 1) Coarse-
grained entity typing methods (Weischedel and
Brunstein,2005;Tokarchuk et al.,2021) assign
mentions to a small set of coarse types; 2) Fine-
grained entity typing (FET) methods (Yuan and
Downey,2018;Onoe et al.,2021) classify men-
tions into more diverse and semantically richer on-
tologies; 3) Ultra-fine entity typing methods (Choi
et al.,2018;Ding et al.,2021;Dai et al.,2021)
use a large open type vocabulary to predict a set of
natural-language phrases as entity types based on
texts. However, FET and ultra-fine entity typing
methods hardly perform satisfactorily due to the
huge predefined type set. They also hardly han-
dle few-shot and zero-shot issues. Comparatively,
our GET model can generate high-quality multi-
granularity types even beyond the predefined set
for the given entity mentions.
Concept Acquisition
Concept acquisition is
very related to entity typing which also aims to
obtain the types for the given entities, since entity
types are often recognized as concepts. Concept
acquisition can be categorized into the extraction-
based or generation-based scheme. The extraction
scheme cannot acquire concepts not existing in the
given text (Yang et al.,2020). The existing ap-
proaches of concept generation (Zeng et al.,2021)
focus on utilizing the existing concept taxonomy or
knowledge bases to generate concepts but neglect
to utilize the large corpus. Our GET model can