Jin R. and Li, X. /Medical Image Analysis (2023) 3
•To the best of our knowledge, we are the first to exam-
ine the robustness of FedGAN, a promising pipeline for
medical data synthesis, from the practical backdoor attack
perspectives. Without loss of generality, we examine our
proposed pipeline, attack, and defense on two public med-
ical datasets.
•We propose a general pipeline of conditional FedGAN to
generate labeled medical datasets. We extend backdoor
attacks for classification models to generative models and
reveal the vulnerability of conditional FedGAN training
by backdoor attacks. We investigate the effect of different
trigger sizes and types in the attacks.
•We propose an effective defense strategy FedDetect and
compare it with the current practices to show its irreplace-
able role in achieving robust FedGANs. We not only
present qualitative results by examining the fidelity of the
synthetic data, but also quantitatively evaluate their util-
ity as data augmentation to assist diagnostic model train-
ing. We show the practical use and the innovation of
FedDetect in the field of medical image synthesis.
A preliminary version of this work, Backdoor Attack is a
Devil in Federated GAN-based Medical Image Synthesis (Jin
and Li, 2022) has been presented in SASHIMI 2022. This paper
extends the preliminary version by expanding the conditional
FedGAN which generates synthetic medical images with labels
so that synthetic images can serve for broader usage (such as
data augmentation), examining the effect of various sizes and
types of triggers with gray-scale and RGB medical datasets,
performing quantitative analysis on the synthetic data, and as-
sessing the utility of the synthetic medical images as data aug-
mentation to assist training deep diagnostic models.
2. Preliminaries and Related Work
2.1. Conditional Generative Adversarial Networks
GAN was first been proposed in Goodfellow et al. (2014) as
an unsupervised generative algorithm, where two deep neural
networks, discriminator and generator are training against each
other to optimize minmax objective function Eq. 1.
min
Gmax
D
Ex∼pdata(x)[log D(x)] +Ez∼pz(z)[log (1 −D(G(z)))],(1)
where Gand Dare generator and discrimnator, xis the training
data, and zis a random noise vector sampled from a predefined
distribution pz.GAN has been used to generate medical im-
age datasets for data augmentation and data sharing for open
research as healthcare institutions are regulated to release their
collected private data (Chen et al., 2022; Lin et al., 2022).
Later, Mirza and Osindero (2014) implemented the condi-
tional GAN, turning GAN to be supervised learning algorithm,
where both the discriminator and the generator take an extra
auxiliary label so that GAN generates images conditional on
the given label according to updated objective function Eq. 2.
min
Gmax
D
Ex∼pdata(x)[log D(x|c)] +Ez∼pz(z)[log (1 −D(G(z|c)))],
(2)
where we add class label cas the conditional term compared
to Eq. 1. Generating synthetic medical data using conditional
GAN is gain more practical values in healthcare scenarios, be-
cause medical data is usually labeled and this label makes it
meaningful for diagnostic purposes, e.g., for classifying if cer-
tain patient has the disease (Frangi et al., 2018), and for data
augmentation (Chen et al., 2022).
2.2. Federated Learning
Training DL models usually requires a large amount of train-
ing data. However, collecting data is challenging in the field of
healthcare because healthcare providers, governments, and re-
lated medical organizations must pay particular attention to the
patient’s privacy and guarantee the proper use of their collected
data (Price and Cohen, 2019). In this case, limited data in local
healthcare institutions is usually biased and unilateral (Wang
et al., 2020b), which in turn impede the AI-assisted diagnostic
technology in healthcare (Van Panhuis et al., 2014).
FL has been proposed as a promising strategy to facilitate
collaboration among several institutions (e.g., medical centers
distributed from different geographical locations) to train a
global DL model (Koneˇ
cn`
y et al., 2016). Given the important
role of FL plays in leveraging medical data from distributed lo-
cations and the practical usage of DL-based synthetic models
in medicine, combining them together will help facilitate ad-
vancement in medicine. Previous studies try to establish the
FedGAN (Rasouli et al., 2020) and explored its robustness in
terms of the differential privacy (Augenstein et al., 2019).
In addition, Byzantine-Robust FL is a key challenge in FL
deployment, as the clients are barely controllable and typically
viewed as an open system. Literature has shown that FL is vul-
nerable to multiple kinds of adversaries (Bouacida and Moha-
patra, 2021; Liu et al., 2022). Example vulnerabilities includes
model poisoning attacks (Bhagoji et al., 2019), gradient inver-
sion attacks (Huang et al., 2021), inference attacks (Ying et al.,
2020), backdoor attacks (Li et al., 2022), etc.
2.3. Backdoor Attack
In this section, we begin by introducing the general concept
of a backdoor attack. Next, we delve into the specific details
of backdoor attacks in FL and backdoor attacks in generative
models.
General concept. The backdoor attackers aim to embed a back-
door, also known as trigger, in the training data in order to cor-
rupt the performance of the Deep Neural Network. Given it in-
volves poisoning data, it belongs to the fields of data poisoning
attacks, which has been widely explored in multiple machine
learning fields, including Support Vector Machines, Statistical
Machine Learning, and DL (Biggio et al., 2012; Nelson et al.,
2008). In DL, current studies mainly explore poisoning attacks
in centralized classification models, where a hidden trigger is
pasted on some of the training data with wrong labels. The
attacker activates it during the testing time so that the classifi-
cation model produces a lower testing accuracy for images with
triggers (Saha et al., 2020). This attack strategy takes advantage
of the tendency that the deep neural network is more likely to