2
ing scenarios. Although several studies have adopted Bayesian
treatments [22] and Gaussian processes [20], [21] to enable
FL methods to handle data uncertainties, their performance
heavily relies on the learning of good posterior inferences and
kernel functions, which is very time-consuming.
To solve the aforementioned issues, in this paper, we
propose a federated fuzzy neural network (FedFNN) with
evolutionary rule learning (ERL) to handle data uncertainties
and non-IID issues. As shown in Fig. 1 (a), the theory of bio-
logical evolution [23] states that variants of the same species
can evolve to adapt to their different living environments by
selectively activating and expressing their genes. Inspired by
this, we use FNNs as our local models and consider them as
compositions of fuzzy rules, which capture valuable local data
information from multiple views, such as distributions. Similar
to the genes of a species, each rule of the FedFNN is a basic
functional component that can be activated or deactivated for
clients according to their performance on local data. Thus,
our FedFNN aims to obtain a group of global fuzzy rules that
can be selectively activated for local clients to enable them
to outperform competing approaches on non-IID data. It is
worth noting that our ERL is a novel algorithm different from
genetic algorithms [24]. These two algorithms are inspired by
the same theory but designed in different ways. The genetic
algorithm treats the whole population as the evolution subject.
It keeps selecting the fittest individuals in each generation
to form more competitive populations until the performance
gets stable. However, it needs to be re-adjusted whenever
the environment is modified. Instead, the ERL considers the
internal diversity of species, which means that a species
normally includes several types of populations (subspecies)
living in diverse environments. It regards subspecies as its
evolution subjects and selectively optimises and shares gene
groups among subspecies until they perform well in different
environments. In our FedFNN, each agent that handles non-IID
data corresponds to a sub-species. The ERL enables agents to
cooperate with each other during their evolutions but preserve
their personalities to adapt to diverse environments. In general,
our FedFNN aims at learning 1) a group of global rules that
capture valuable information among local clients and 2) a
rule activation strategy for each local client to ensure the
personalization and superior performance of the FedFNN.
The ERL is an iterative learning approach with two stages:
a rule cooperation stage, where the global rules are updated
cooperatively by clients based on their rule activation statuses,
and a rule evolution stage, where the activation statuses of
local rules are adjusted according to their contributions to
the performance of the FedFNN on non-IID local data. The
former stage enhances the cooperation among clients for
learning more representative global rules, which increases the
generalizability of the FedFNN. In contrast, the latter stage
fulfils the selective activation of rules that enable the local
FNNs to be more adaptive and perform better on non-IID
data, which improves the personalization of the FedFNN. For
an explanation of the FedFNN model, refer to the diagram
shown in Fig. 1 (b).
The contributions of this paper are as follows,
•We are the first to propose FedFNN that integrates fuzzy
neural networks into a federated learning framework to
handle data non-IID issues as well as data uncertainties
in distributed scenarios. FedFNN is able to learn person-
alized fuzzy if-then rules for local clients according to
their non-IID data.
•Inspired by the theory of biological evolution, we design
an ERL algorithm for the learning of FedFNN. ERL
encourages the global rules to evolve while selectively
activating superior rules and eliminating inferior ones
during the training procedure. This procedure ensures the
ability of generalization and personalization of FedFNN.
II. RELATED WORK
A. Distributed fuzzy neural networks
DFNNs [8]–[12] have been proposed to handle the uncer-
tainties encountered in distributed applications. The authors in
[8] proposed a DFNN model that randomly sets the parameters
in antecedents and only updates the parameters in consequent
layers. Later, they extended this work to an online DFNN
model [9]. Their models assume that all clients share the
information in antecedent layers, making this technically not
a seriously distributed method. To avoid this problem, a
fully DFNN [10] model was proposed by adopting consensus
learning in both the antecedent and consequent layers. As
its subsequence variant, a semisupervised DFNN model [12]
was presented to enable the DFNN to leverage unlabeled
samples by using the fuzzy C-means method and distributed
interpolation-based consistency regularization. However, the
existing DFNNs cannot handle situations in which the data dis-
tribution varies across clients. The authors in [11] proposed a
DFNN with hierarchical structures to process the heterogeneity
existing in the variables of training samples. However, instead
of processing the data heterogeneity across distributed clients,
they focused on variable composition heterogeneity, which
meant that data variables were collected from different sources.
Generally, by employing the well-known Takagi-Sugeno (T-
S) fuzzy if-then rules [2], the existing DFNN models build
the antecedent layers of their local models in traditional ways
(e.g., K-means) and calculate the corresponding consequent
layers with closed-form solutions. Then, the original DFNNs
are transformed into convex optimization problems. While
efficient and effective, they are not able to learn local models
with personalized rule sets. Worse, they fail to utilize the
strong learning abilities of neural networks that enable local
FNNs to investigate more adaptive rules.
B. Federated Learning
FL [14] is an emerging distributed paradigm in which
multiple clients cooperatively train a neural network without
revealing their local data. Recently, many solutions [14], [25],
[26] have been presented to solve FL problems, among which
the most known and basic solution is federated averaging
(FedAvg) [14], which aggregates local models by calculating
the weighted average of their updated parameters.
However, FL has encountered various challenges [27],
among which the non-IID issue is the core problem that
makes the local model aggregation process harder and leads