applications, PLL is usually in a scenario where the label-
ing resources are constrained, and the adequacy of partial
annotation is not guaranteed. Under this circumstance, ex-
isting consistency regularization-based methods often fails
to achieve satisfactory performances. As shown in Figure
1, DPLL (Wu, Wang, and Zhang 2022) and PiCO (Wang
et al. 2022b) achieve state-of-the-art performances when us-
ing complete partial training set, but as the proportion of par-
tial example decreases, their accuracies drop significantly.
The reason behind this phenomenon is that when the num-
ber of labels is scarce and inherently ambiguous, there is
not enough supervision information to guide the initial su-
pervised learning of the model, which leads to the conver-
gence of the consistency regularization to the wrong direc-
tion and the emergence of problems such as overfitting and
class-imbalance.
Witnessing the enormous power of unlabeled examples
via consistency regularization (Sohn et al. 2020; Berthelot
et al. 2020; Xie et al. 2020), we hope to facilitate par-
tial label consistency regularization through these readily
available data. To this end, an effective framework needs to
be designed to maximize the potential of partial and unla-
beled examples, as well as reasonable mechanisms to guide
the model when supervision information is scarce and am-
biguous. In this paper, we propose consistency regulariza-
tion with controller (abbreviated as ConCont). Our method
learns from the supervised information in the training targets
(i.e., candidate label sets) via a supervised loss, while per-
forming controller-guided consistency regularization at both
label- and representation-levels with the help of unlabeled
data. To avoid negative regularization, the controller divides
the examples as confident or unconfident according to the
prior information and the learning state of the model, and
applies different label- and representation-level consistency
regularization strategies, respectively. Furthermore, we dy-
namically adjust the confidence thresholds so that the num-
ber of samples of each class participating in consistency reg-
ularization remains roughly equal to alleviate the problem of
class-imbalance.
Related Work
Traditional PLL methods can be divided into two categories:
averaging-based (H¨
ullermeier and Beringer 2006; Cour,
Sapp, and Taskar 2011; Zhang, Zhou, and Liu 2016) and
identification-based (Jin and Ghahramani 2002; Liu and Di-
etterich 2012; Feng and An 2019; Ni et al. 2021). Averaging-
based methods treat all the candidate labels equally, while
identification-based methods aim at identifying the ground-
truth label directly from candidate label set. With the pop-
ularity of deep neural networks, PLL has been increasingly
studied in deep learning paradigm. Yao et al. (2020) made
the first attempt with an entropy-based regularizer enhanc-
ing discrimination. Lv et al. (2020) propose a classifier-
consistent risk estimator for partial examples that theoreti-
cally converges to the optimal point learned from its fully su-
pervised counterpart under mild condition, as well as an ef-
fective method progressively identifying ground-truth labels
from the candidate sets. Wen et al. (2021) propose a family
of loss functions named leveraged weighted loss taking the
trade-offs between losses on partial labels and non-partial
labels into consideration, advancing the former method to a
more generalized case. Xu et al. (2021) consider the learning
case where the candidate labels are generated in an instance-
dependent manner.
Recently, consistency regularization-based PLL methods
have achieved impressive results, among which two repre-
sentatives are: PiCO (Wang et al. 2022b) and DPLL (Wu,
Wang, and Zhang 2022). They can be seen as perform-
ing consistency regularization at the representation-level and
label-level, respectively. To be specific, PiCO aligns the rep-
resentations of the augmented variants of samples belonging
to the same class, and calculates a representation prototype
for each class, and then disambiguates the label distribution
of the sample according to the distance between the sample
representation and each class prototype, forming an itera-
tive EM-like optimization. While DPLL aligns the output la-
bel distributions of multiple augmented variants to a confor-
mal distribution, which serves as a comprehensiveness of the
label distribution for all augmentations. Despite achieving
state-of-the-art performances under fully PLL datasets, their
consistency regularization rely heavily on the sufficiency of
partial annotations, which greatly limits their applications.
Our work is also related with semi-supervised PLL
(Wang, Li, and Zhou 2019; Wang and Zhang 2020). Despite
similar learning scenarios, previous semi-supervised PLL
methods are all based on nearest-neighbor or linear classi-
fiers, and have not been integrated with modern consistency
regularized deep learning, which are very different with our
method in terms of algorithm implementation.
Methodology
Notations
Let X ⊆ Rdbe the input feature space and Y=
{1,2, . . . , C}denote the label space. We attempt to induce
a multi-class classifier f:X 7→ [0,1]Cfrom partial la-
bel training set Dp={(xi,yi)|1≤i≤p}and an ad-
ditional unlabeled set Du={xi|p+ 1 ≤i≤p+u}.
Here, yi= (yi
1, yi
2, . . . , yi
C)∈ {0,1}Cis the partial label,
in which yi
j= 1 indicates the j-th label belonging to the
candidate set of sample xiand vice versa. Following the
basic assumption of PLL, the ground-truth label ℓ∈ Y is
inaccessible to the model while satisfies yi
ℓ= 1. In order
to facilitate the unified operation of partial and unlabeled
examples, the target vector of unlabeled examples is repre-
sented as (1,1,...,1), i.e., the candidate set equals to Y,
containing no label information. For the classifier f, we use
fj(x)to denote the output of classifier fon label jgiven
input x.
Consistency Regularization with Controller
Briefly, our method learns from the supervised information
in the training targets (i.e., candidate label sets) via a super-
vised loss, while performing controller-guided consistency
regularization at both label- and representation-level with
the help of unlabeled data. The consistency regularization
here works by aligning the outputs of different augmented
variants of each example. To prevent the model from falling