
Figure 1. Examples of synthetic face images in our dataset. Our dataset captures a wide variety of facial geometry, pose, textures, expres-
sions, accessories and environments.
son; and (3) the same person that goes by different names
labeled as different persons.
(3) Data bias. Face recognition models are generally
trained and tested on celebrity faces, many of which
are taken with strong lighting and make-up. Celebrity
faces also have imbalanced racial distribution (e.g., 84.5%
of the faces in CASIA-WebFace [39] are Caucasian
faces [34]), leading to poor recognition accuracy for the
under-represented racial groups [34].
In order to circumvent all these issues that affect the ex-
isting real face datasets, we introduce a new large-scale face
recognition dataset consisting only of photo-realistic digital
face images rendered using a computer graphics pipeline
and make this dataset available to the community. Specifi-
cally, we build upon the face generation pipeline introduced
by Wood et al. [36], tailoring the amount of variability for
each attribute (e.g., pose and accessories) for our recogni-
tion task, and generate 1.22M images with 110K unique
identities. Each identity is generated by randomizing the
facial geometry and texture as well as the hair style. The
generated face is then rendered with different poses, ex-
pressions, hair color, hair thickness and density, accessories
(including clothes, make-ups, glasses, and head/face wear),
cameras and environments, to encourage the network to
learn a robust embedding. Figure 1 shows examples of syn-
thetic face images in this new dataset. We generated 1.22M
images, but in practice the number of identities and images
you can generate with synthetics pipeline is only limited by
the cost of generating and storing these images.
Digital synthetic faces can solve the aforementioned
problems associated with the real face datasets. Firstly, the
generated faces are free of label noise. Secondly, the bias in
lighting, make-up and skin color can be reduced as we have
full control over those attributes. Most importantly, the face
generation pipeline does not rely on any privacy-sensitive
data obtained without consent.
This is a critical difference from the GAN-generated syn-
thetic faces; face GANs rely (either directly or indirectly) on
large-scale real face datasets to train some components of
their pipeline, leaving unresolved ethical problems. For ex-
ample, a recent method called SynFace [28] was trained on
synthetic faces generated using DiscoFaceGAN [9]. While
the generated face images are free of label noise, millions
of real face images were used for training DiscoFaceGAN.
The GANs may also inherit any bias that exists in the real
face images used to train them. For our dataset, only 511
face scans, obtained with consent, were used to build a
parametric model of face geometry and texture library [36].
From this limited source data, we can generate infinite num-
ber of identities, making our approach easily scalable.
Our contributions can be summarized as below:
• We release a new large-scale synthetic dataset for face
recognition that is free from privacy violations and lack
of consent. To the best of our knowledge, our dataset,
containing 1.22M images of 110K identities, is the largest
public synthetic dataset for face recognition.
• Compared to SynFace [28], which is trained on GAN-
generated faces, we reduce the error rate on LFW by
52.5% (accuracy from 91.93% to 96.17%). For five popu-
lar benchmarks [14, 30, 41, 25, 42], the average error rate
is reduced by 46.0% (accuracy from 74.75% to 86.37%).
• We demonstrate how the proposed synthetic dataset can
be used in conjunction with a small number of real face
images to substantially improve the accuracy. This sim-
ulates a scenario where a small number of curated (i.e.,
no label noise and reduced bias) real face images are col-
lected with consent. By fine-tuning our network with only
120K real face images (i.e., 2% of the commonly-used
MS1MV2 dataset [8]), we achieve 99.33% accuracy on
LFW and 93.61% on average across the five benchmarks,
which is comparable to the methods trained on millions
of real face images.
• Having full control over the rendering pipeline, we per-
form extensive experiments to study how each attribute
(e.g., variation in facial pose, accessories and textures)
affects the face recognition accuracy.