deepfake detection model, there exists an implicit decision
boundary between fake images and genuine images based
on identities. During the training phase, binary classifiers
may accidentally consider certain groups of identities as
genuine identities and other groups of identities as fake
identities. When tested on the cross-dataset evaluation, such
biased representations may be mistakenly used by binary
classifiers, causing false judgments based on the facial ap-
pearance of images. In this paper, we have qualitatively
and quantitatively verified this phenomenon (termed as the
Implicit Identity Leakage) in binary classifiers of various
backbones. Please see Sec. 3and Sec. 5.2 for analyses.
Furthermore, based on such understanding, we propose
a simple yet effective method named the ID-unaware Deep-
fake Detection Model to reduce the influence of Implicit
Identity Leakage. Intuitively, by forcing models to only fo-
cus on local areas of images, less attention will be paid to
the global identity information. Therefore, we design an
anchor-based detector module termed as the Artifact Detec-
tion Module to guide our model to focus on the local artifact
areas. Such a module is expected to detect artifact areas on
images with multi-scale anchors, each of which is assigned
a binary label to indicate whether the artifact exists. By lo-
calizing artifact areas and classifying multi-scale anchors,
our model learns to distinguish the differences between lo-
cal artifact areas and local genuine areas at a finer level, thus
reducing the misusage of the global identity information.
Extensive experimental results show that our model
accurately predicted the position of artifact areas and
learned generalized artifact features in face manipulation
algorithms, successfully outperforming the state-of-the-art.
Contributions of the paper are summarized as follows:
• We discover that deepfake detection models super-
vised only by binary labels are very sensitive to the
identity information of the images, which is termed as
the Implicit Identity Leakage in this paper.
• We propose a simple yet effective method termed as
the ID-unaware Deepfake Detection Model to reduce
the influence of the ID representation, successfully
outperforming other state-of-the-art methods.
• We conduct extensive experiments to verify the Im-
plicit Identity Leakage phenomenon and demonstrate
the effectiveness of our method.
2. Related Work
With the development of Generative Adversarial Net-
work (GAN) [12,25,33,34,47] techniques, forgery im-
ages/videos have become more realistic and indistinguish-
able. To deal with attacks based on different face manip-
ulation algorithms, researchers tried to improve their deep-
fake detectors [30,54,61] from different perspectives, such
as designing different loss functions [7], extracting richer
features [21,83], and analyzing the continuity between con-
secutive frames [29,57]. Most of these deepfake detection
methods can be roughly summarized into two categories.
2.1. Binary Classifiers
Many researchers [1,6,14,56,62] treated the deepfake
detection task as a binary classification problem. They used
a backbone encoder to extract high-level features and a clas-
sifier to detect whether the input image has been manip-
ulated. Durall et al. [22] first proposed a model analyz-
ing the frequency domain for face forgery detection. Masi
et al. [52] used a two-branch recurrent network to extract
high-level semantic information in original RGB images
and their frequency domains at the same time, by which
the model achieved good performance on multiple public
datasets. Li et al. [39] designed a single-center loss to
compress the real sample classification space to further im-
prove the detection rate of forged samples. Binary classi-
fiers achieved high detection accuracy on in-dataset evalu-
ation, but they could not maintain good performance when
facing unseen forged images.
2.2. Hand-crafted Deepfake Detectors
Many works attempted to improve the generalization ca-
pability of deepfake detectors by modeling specific hand-
crafted artifacts among different face manipulation meth-
ods. Li et al. [42] believed that some physical characteris-
tics of a real person cannot be manipulated in fake videos.
They designed an eye blinking detector to identify the au-
thenticity of the video through the frequency of eye blink-
ing. Since 3D data can not be reversely generated from the
fake image, Yang et al. [84] did the face forgery detection
task from the perspective of non-3D projection generation
samples. Sun et al. [74] and Li et al. [41] focused on pre-
cise geometric features (face landmark) and blending ar-
tifacts respectively when detecting forged images. Liu et
al. [46] equipped the model with frequency domain infor-
mation since the frequency domain is very sensitive to up-
sampling operations (which are often used in deepfake de-
tection models), and used a shallow network to extract rich
local texture information, enhancing the model’s general-
ization and robustness.
In summary, hand-crafted deepfake detectors guided the
model to capture specific artifact features and indicated ma-
nipulated images/videos by responding to these features.
However, these methods have a common limitation: when
forgeries do not contain specific artifacts that are introduced
in the training phase, they often fail to work well.
3. Implicit Identity Leakage
The Implicit Identity Leakage denotes that the ID repre-
sentation in the deepfake dataset is captured by binary clas-
2