1
Over-the-Air Federated Learning with Privacy
Protection via Correlated Additive Perturbations
Jialing Liao, Zheng Chen, and Erik G. Larsson
Department of Electrical Engineering (ISY), Link¨
oping University, Link¨
oping, Sweden
Email: {jialing.liao, zheng.chen, erik.g.larsson}@liu.se
Abstract—In this paper, we consider privacy aspects of wireless
federated learning (FL) with Over-the-Air (OtA) transmission
of gradient updates from multiple users/agents to an edge
server. OtA FL enables the users to transmit their updates
simultaneously with linear processing techniques, which improves
resource efficiency. However, this setting is vulnerable to privacy
leakage since an adversary node can hear directly the un-
coded message. Traditional perturbation-based methods provide
privacy protection while sacrificing the training accuracy due
to the reduced signal-to-noise ratio. In this work, we aim at
minimizing privacy leakage to the adversary and the degradation
of model accuracy at the edge server at the same time. More
explicitly, spatially correlated perturbations are added to the
gradient vectors at the users before transmission. Using the
zero-sum property of the correlated perturbations, the side
effect of the added perturbation on the aggregated gradients
at the edge server can be minimized. In the meanwhile, the
added perturbation will not be canceled out at the adversary,
which prevents privacy leakage. Theoretical analysis of the
perturbation covariance matrix, differential privacy, and model
convergence is provided, based on which an optimization problem
is formulated to jointly design the covariance matrix and the
power scaling factor to balance between privacy protection
and convergence performance. Simulation results validate the
correlated perturbation approach can provide strong defense
ability while guaranteeing high learning accuracy.
I. INTRODUCTION
As one instance of distributed machine learning, federated
learning (FL) was developed by Google in 2016, where the
clients can train a model collaboratively by exchanging local
gradients or parameters instead of raw data [1]. Research
activities on FL over wireless networks have attracted wide
attention from various perspectives, such as communication
and energy efficiency, privacy and security issues etc [2], [3].
Communication efficiency is an important design aspect of
wireless FL schemes due to the need of data aggregation over
a large set of distributed nodes with limited communication
resources. Recently, Over-the-Air (OtA) computation has been
applied for model aggregation in wireless FL by exploiting the
waveform superposition property of multiple-access channels
[4], [5]. Under OtA FL, edge devices can transmit local gra-
dients or parameters simultaneously, which is more resource-
efficient than traditional orthogonal multiple access schemes.
Despite the extensive research on wireless FL, recent works
have shown that traditional FL schemes are still vulnerable to
inference attacks on local updates to recover local training
data [6], [7]. One solution is to reduce information disclosure,
This work is supported by Security Link, ELLIIT, and the KAW foundation.
which motivates the usage of compression methods such as
dropout, selective gradients sharing, and dimensionality reduc-
tion [8]–[10], with the drawbacks of limited defense ability and
no accuracy guarantee. Other cryptography technologies, such
as secure multi-party computation and homomorphic encryp-
tion [11], [12] can provide strong privacy guarantees, but yield
more computation and communication costs while being hard
to implement in practice. Due to easy implementation and high
efficiency, perturbation methods such as differential privacy
(DP) [13] or CountSketch matrix [14] have been developed.
DP technique can effectively quantify the difference in output
caused by the change in individual data and reduce information
disclosure by adding noise that follows some distributions
(e.g., Gaussian, Laplacian, Binomial) [13], [15]. In the context
of FL, one can use two DP variants by transmitting perturbed
local updates or global updates, i.e., Local DP and Central
DP [16]. However, DP-based methods fail to achieve high
learning accuracy and defense ability at the same time due to
the reduction of signal-to-noise ratio (SNR), which ultimately
limits their application.
To address this issue, in this paper, we design an efficient
perturbation method for OtA FL with strong defense ability
without significantly compromising the learning accuracy. Un-
like the traditional DP method by adding uncorrelated noise,
we add spatially correlated perturbations to local updates at
different users/agents. We let the perturbations from different
users sum to zero at the edge server such that the learning
accuracy is not compromised (with only slightly decreased
SNR due to less power for actual data transmission). On the
other hand, the perturbations still exist at the adversary due
to the misalignment between the intended channel and the
eavesdropping channel, which can prevent privacy leakage.
A. Related Work
The authors in [17] developed a hybrid privacy-preserving
FL scheme by adding perturbations to both local gradients and
model updates to defend against inference attacks. In [18] the
client anonymity in OtA FL was exploited by randomly sam-
pling the devices participating and distributing the perturbation
generation across clients to ensure privacy resilience against
the failure of clients. Without adversaries but with a curious
server, the trade-offs between learning accuracy, privacy, and
wireless resources were discussed in [19]. Later on, authors of
[20] developed a privacy-preserving FL scheme under orthog-
onal multiple access (OMA) and OtA, respectively, proving
arXiv:2210.02235v1 [cs.LG] 5 Oct 2022