However, efficient model recovery in FL is largely un-
explored. Since the server does not know in which round
the attack happens, the server may not be able to simply
roll back to a clean global model in a prior round. A naive
recovery method (we call it train-from-scratch) is to remove
the detected malicious clients and train a new global model
from scratch using the remaining clients. Train-from-scratch
could recover an accurate global model. However, it introduces
substantial computation and communication cost to the clients
since it requires them to participate in the entire training
process once again. Such computation and communication cost
may be intolerable for resource-constrained clients such as
smartphones and IoT devices.
Our work: In this work, we propose FedRecover, a method
that can recover an accurate global model from a poisoned one
while introducing small computation and communication cost
for the clients. Like train-from-scratch, FedRecover removes
the detected malicious clients, re-initializes a global model,
and trains it iteratively in multiple rounds. However, unlike
train-from-scratch, FedRecover reduces the cost for the clients
by changing the way of obtaining their model updates. Our
intuition is that the historical information, including the global
models and clients’ model updates, which the server collected
when training the poisoned global model before the malicious
clients are detected, still carry valuable information for model
recovery. Based on the intuition, our key idea is that, during the
recovery process, the server estimates the remaining clients’
model updates using such historical information instead of ask-
ing the clients to compute and communicate them. FedRecover
is independent of the detection methods used to detect the
malicious clients and the aggregation rules of FL. In other
words, FedRecover can be used together with any detection
method and FL aggregation rule in a defense-in-depth strategy.
The key of FedRecover is that the server estimates the
clients’ model updates itself during the recovery process.
Specifically, the server stores the historical information when
training the poisoned global model before the malicious clients
are detected. During the recovery process, the server uses
the well-known Cauchy mean value theorem to estimate each
client’s model update in each round. However, the Cauchy
mean value theorem requires an integrated Hessian matrix for
each client, whose exact value is challenging to compute. To
address the challenge, we further leverage an L-BFGS based
algorithm to efficiently approximate the integrated Hessian
matrix. FedRecover introduces some storage and computation
cost to the server due to storing the historical information and
estimating the clients’ model updates. However, such cost is
acceptable since the server is powerful.
Since FedRecover estimates the clients’ model updates, the
estimation errors may accumulate over multiple rounds during
the recovery process, which eventually may result in a less ac-
curate recovered global model. We propose multiple strategies
to address the challenge. Specifically, the L-BFGS algorithm
requires the recovered global models in the previous several
rounds to estimate a client’s model update in the current round.
The accurately recovered global models in the first several
rounds of the recovery process will help reduce the estimation
errors in the future rounds. Therefore, we propose the warm-
up strategy, in which the server asks the clients to compute and
communicate their exact model updates in the first Twrounds
of the recovery process. Moreover, we propose the periodic
correction strategy, in which the server asks the clients to
compute and communicate their exact model updates in every
Tcrounds. When an estimated model update for a client is
large, it has large influence on the recovered global model. To
reduce the impact of potentially incorrectly estimated large
model updates, we propose the abnormality fixing strategy,
in which the server asks a client to compute its exact model
update when at least one coordinate of the estimated model
update is larger than a threshold τ. Furthermore, we propose
final tuning strategy to reduce the estimation error before the
training terminates, in which the server asks the clients to
compute and communicate their exact model updates in the
last Tfrounds. The parameters Tw,Tc,τ, and Tfcontrol the
trade-off between accuracy of the recovered global model and
computation/communication cost for the clients. In particular,
a larger Tw, a smaller Tc, a smaller τ, or a larger Tfmay
recover a more accurate global model but also introduces a
larger cost to the clients.
Theoretically, we show that the difference between the
global model recovered by FedRecover and the global model
recovered by train-from-scratch can be bounded under some
assumptions, e.g., the loss function used to learn the global
model is smooth and strongly convex. Empirically, we evaluate
FedRecover extensively using four datasets, three FL methods
(e.g., FedAvg [24], Median [36], and Trimmed-mean [36]),
as well as Trim attack (an untargeted poisoning attack) [18]
and backdoor attack (a targeted poisoning attack) [5]. Our
empirical results show that FedRecover can recover global
models that are as accurate as those recovered by train-from-
scratch while saving lots of computation/communication cost
for the clients. For instance, the backdoor attack with 40
malicious clients can achieve 1.00 attack success rate when the
dataset is MNIST and the FL method is Trimmed-mean. Both
FedRecover and train-from-scratch can recover global models
with 0.07 test error rate and 0.01 attack success rate, but Fe-
dRecover saves the clients’ computation/communication cost
by 88% on average compared to train-from-scratch. Moreover,
FedRecover can efficiently recover as accurate global models
as train-from-scratch even if the detection method incorrectly
detects some malicious clients as benign and/or some benign
clients as malicious.
In summary, our key contributions are as follows:
•We perform the first systematic study on model recovery
from poisoning attacks in FL.
•We propose FedRecover to recover a global model via
estimating clients’ model updates through historical in-
formation and multiple optimization strategies.
•We evaluate FedRecover both theoretically and empiri-
cally. Our results show that FedRecover can recover a
global model both accurately and efficiently.