construct the estimator, which enjoys a high efficiency for
detection, especially when there are large-scale clients in
federated learning. We then propose using a small public
dataset (i.e., less than 10 samples) to train an initial global
model, which is crucial for improving the detection accu-
racy.
In summary, we make the following contributions:
•We propose a new Byzantine-robust federated learn-
ing scheme called Robust-FL. To the best of our
knowledge, Robust-FL is the first predication-based
defense scheme that can mitigate Byzantine attacks
effectively and efficiently without relying on any
fallacious assumptions.
•We propose incorporating clustering algorithms to
adaptively adjust the differences between the estima-
tor and local models, such that a boundary between
benign and malicious models can be effectively af-
firmed to identify Byzantine participants.
•We conduct extensive experiments to evaluate
Robust-FL. The results show that Robust-FL is still
effective even more than 50% participants are com-
promised, the Byzantine models are variable, and the
participants’ data are not available, while all existing
defenses are invalid under this severe scenario.
2. Related Work
In order to resist Byzantine attacks, researchers have
proposed many defensive schemes in recent years. We divide
them into three categories according to the principles that the
server relied on to detect or evade anomalous local models.
Distance-based defenses: The first category focuses on
comparing the distances between the local models to find
out anomalous ones. Krum [2] aims to choose one local
model that is closest to its K−f−2neighbors, where
Kis the number of participants and fis the number of
malicious users. Since Krum converges slowly, the authors
introduced its variant Multi-Krum, which chooses K−f
local models for aggregation rather than just one. Similar
to Multi-Krum, FABA [23] iteratively removes the local
model that is farthest from the average model until the
number of eliminated models is f. FoolsGold [7] uses cosine
similarity to identify malicious models and then assigns
them smaller weights to reduce their impact on the averaged
global model. Sniper [3] selects local models for aggregation
based on a graph which is constructed according to the
Euclidean distances between the local models. The PCA
scheme [20] projects local updates into two-dimensional
space and uses a clustering algorithm to find malicious
updates. MAB-RFL [21] is also equipped with PCA and
clustering algorithm to identify malicious updates, in ad-
dition, a momentum based approach is applied to tackle
the data heterogeneity (i.e., Non-IID) challenge. All these
solutions (except MAB-RFL), however, only work well over
independently and identically distributed (IID) data, and
they cannot tolerate more than 50% attackers. Besides, most
of them need to know the number of attackers in advance.
Statistics-based defenses: The second category exploits
the statistical characteristics to remove statistical outliners.
Instead of performing detection-then-aggregation, Trimmed
Mean [30] directly uses all the local updates to obtain a
new global model, by computing the median or the trimmed
mean of all local models in each dimension. Geometric
Median [25] intends to find a new update that minimizes
the summation of the distances between the update and each
local model. The RFA scheme [17] computes the geometric
median of the local models with an alternating minimization
approach to reduce the computational overhead. Bulyan [16]
first uses Multi-Krum to remove malicious models and
then aggregates the rest models based on Trimmed Mean.
SLSGD [26] also adopts Trimmed Mean as the aggrega-
tion rule, and then uses a newly proposed moving-average
method, which considers global models in this round and the
last round. Nevertheless, the above schemes cannot identify
Byzantine users, and they perform poorly when there are
more than 50% Byzantine users.
Performance-based defenses: The last category de-
pends on the validation dataset to evaluate the performance
of the uploaded parameters. Li et al. [14] proposed us-
ing a pre-trained autoencoder to detect malicious models.
Zeno [27] computes the stochastic descendant score for each
gradient based on a validation dataset and then removes
the gradients with low stochastic descendant scores. Cao
et al. [5] proposed a Byzantine-robust distributed gradient
algorithm, which computes a noisy gradient based on a clean
dataset, and a gradient is accepted only when its distance
between the noisy gradient satisfies a pre-defined condition.
Prakash et al. [18] proposed DiverseFL, which first com-
putes a guiding gradient for each user based on the data
the user shares, and then two similarity metrics (Direction
Similarity and Length Similarity) between the local gradient
and the corresponding guiding gradient are considered, only
when both metrics are satisfied will the gradient be accepted.
FLTrust [4] bootstraps trust with a clean training dataset
collected by the server. More specifically, the RELU-clipped
cosine-similarity between each local update and the server
update (calculated on the cleaning dataset) is employed to
reweight the local update, such that malicious updates have
a limited impact on the global model. However, algorithm
[14] requires a lot of data to obtain benign models and trains
autoencoder based on the benign models, but in reality, it is
hard to obtain so much data. Although the rest four schemes
require few data, they have other limitations. For instance,
Zeno needs to know the number of attackers in advances;
scheme [5] relies on an appropriate hyper-parameter to dis-
tinguish benign gradients from malicious ones; DiverseFL
compels users to share their private data, which violates the
original intention of FL; FLTrust cannot identify Byzantine
users, which means that malicious updates can also partici-
pate in aggregation to deteriorate the accuracy of the global
model.