Learning Real Degradation from Face to Natural Images 3
complex real degradation which cannot be well synthesized (see (d) and (e) in
Figure 1). By incorporating the synthetic halftone degradation [13], BSRGAN*
has slight improvement (see (f)), but still contains obvious linearity artifacts.
In contrast, face image has specific and strong structure prior, and can be
better restored while exhibiting great generalization ability on real-world LQ
images in most cases [27, 47, 53]. Although the image is corrupted by intractable
degradation, the face restoration result is very plausible and photo-realistic (see
Figure 1 (h)). Since the face and non-face (natural) regions in an image share the
same degradation, once we have known the degradation process on face regions,
transferring it to natural HQ images would bring considerable benefits, e.g., we
can apply this degradation process on the HQ natural image to synthesize these
types of natural image pairs (see (i)) for training restoration network (see (g)).
In this paper, we make the first attempt to explore the
re
al
deg
radation with
ReDegNet, which contains (i) learning the real degradation from the pairs of
real-world LQ and pseudo HQ face images with DegNet, and (ii) transferring it
to HQ natural images to synthesizing their realistic LQ ones with SynNet. As for
(i), instead of taking a single LQ image to predict its degradation parameters [19],
our DegNet takes the real-world LQ and its pseudo HQ face images as input to
generate the degradation representation, which models the degradation process
of how the HQ image is degraded to the LQ one. To disentangle the image
content and degradation type, we adopt two manners, i.e., a) carefully designed
framework by predicting the degradation representation through several fully
connected layers to generate the convolution weights which can be regarded as
the styles in StyleGANs [22, 23], and b) contrastive loss [46] by minimizing the
representation distance between the pairs with different content but degraded
with the same degradation parameters, and meanwhile maximizing these with
the same content but different degradation. This process is fully supervised by
the paired LQ/HQ face images. As for (ii), our SynNet synthesizes the realistic
LQ natural images with these degradation representations extracted from face
images, which can help us to learn the real-world restoration mapping. Note that
our method may perform limited on scenarios without faces. By extending the
degradation space with face images share the similar degradation, our model
would be further improved. The main contributions are summarized as follows:
–
We propose the ReDegNet to explore the real degradation from face im-
ages by explicitly learning the degradation-aware and content-independent
representations which control the degraded image generation.
–
We transfer these real-world degradation representations to HQ natural
images to generate their realistic LQ ones for supervised real restoration.
–
We provide a new manner for handling intractable degraded images by
learning their degradation from face regions within them, which can be used
for synthesizing these types of LQ natural images for specifically fine-tuning.
–
Experimental results demonstrate that our ReDegNet can well learn the
degradation representations from face images and can effectively transfer
to natural ones, contributing to the comparable performance on general
restoration and superior performance in specific scenarios against the SOTAs.