FEAMOE Fair Explainable and Adaptive Mixture of Experts Shubham Sharma

2025-04-27 0 0 1.04MB 20 页 10玖币
侵权投诉
FEAMOE: Fair, Explainable and Adaptive Mixture
of Experts
Shubham Sharma
University of Texas at Austin
shubham_sharma@utexas.edu
Jette Henderson
CognitiveScale
jhenderson@cognitivescale.com
Joydeep Ghosh
University of Texas at Austin
jghosh@utexas.edu
Abstract
Three key properties that are desired of trustworthy machine learning models
deployed in high-stakes environments are fairness, explainability, and an ability to
account for various kinds of "drift". While drifts in model accuracy, for example
due to covariate shift, have been widely investigated, drifts in fairness metrics
over time remain largely unexplored. In this paper, we propose FEAMOE, a
novel "mixture-of-experts" inspired framework aimed at learning fairer, more
explainable/interpretable models that can also rapidly adjust to drifts in both the
accuracy and the fairness of a classifier. We illustrate our framework for three
popular fairness measures and demonstrate how drift can be handled with respect
to these fairness constraints. Experiments on multiple datasets show that our
framework as applied to a mixture of linear experts is able to perform comparably
to neural networks in terms of accuracy while producing fairer models. We then
use the large-scale HMDA dataset and show that while various models trained on
HMDA demonstrate drift with respect to both accuracy and fairness, FEAMOE
can ably handle these drifts with respect to all the considered fairness measures
and maintain model accuracy as well. We also prove that the proposed framework
allows for producing fast Shapley value explanations, which makes computationally
efficient feature attribution based explanations of model decisions readily available
via FEAMOE.
1 Introduction
The field of responsible artificial intelligence has several desiderata that are motivated by regulations
such as the General Data Protected Regulation [
9
]. These include: ensuring that an AI model is
non-discriminatory and transparent; individuals subject to model decisions should have access to
explanations that point a path towards recourse; and models should adapt to any changes in the
characteristics of the data post-deployment so as to maintain their quality and trustworthiness.
Most approaches towards the mitigation of any form of bias assume a static classifier. A practitioner
decides on some definition of fairness, trains a model that attempts to enforce this notion of fairness
and then deploys the model. Many of the fairness definitions are based on model outcomes or
on error rates (the gap between true and/or false positive rates) that are associated with different
subgroups specified by a protected attribute. The goal is to reduce the difference between these error
rates across relevant subgroups. For example, average odds difference [
4
] is a measure signifying
equalized odds and is given by the sum of the differences in both true positive and false positive rates
between two groups, scaled by a factor of 0.5. Equality of opportunity and demographic parity [
3
]
Preprint. Under review.
arXiv:2210.04995v1 [cs.LG] 10 Oct 2022
are also popular definitions of fairness. Recently, fairness in terms of a gap of recourse has been
proposed, where recourse is defined as the ability to obtain a positive outcome from the model [
34
].
While the suitability of a fairness measure is application dependent [
26
,
3
], demographic parity and
equalized odds remain the most popularly used, and the need for recourse gap-based fairness is being
increasingly recognized [18].
However, static models can encounter drift once deployed, as the statistical properties of real data
often change over time. This can lead to deteriorating performance. Model drift can occur when the
properties of the target variable change (concept drift) or when the input data distribution changes, or
both. The performance of models has largely been measured through accuracy-based metrics such
as misclassification rates, F-score or AUC. [
37
]. However, a model trained in the past and found
to be fair at training time may act unfairly for data in the present. Addressing drift with respect to
fairness in addition to accuracy has remained largely unexplored though it is an important aspect of
trustworthy AI in practice.
Explainability of individual model outcomes is another principal concern for trustworthy ML. Among
many methods of explanations in terms of feature attribution, [
6
], the SHAP approach based on Shap-
ley values is particularly popular as it enjoys several axiomatic guarantees [21]. While computation
of SHAP values is fast for linear and tree-based models, it can be very slow for neural networks
and several other model types, especially when the data has a large numbers of features or when a
large number of explanations are required [
27
]. This poses a barrier to deployments that demand fast
explanations in real-time, production settings.
In this paper, we address these fairness, data/model drift, and explainability concerns by proposing
FEAMOE: Fair, Explainable and Adaptive Mixture of Experts, an incrementally grown mixture of
experts (MOE) with fairness constraints. In the standard mixture of experts setup, each expert is
a machine learning model, and so is the gating network. The gating network learns to assign an
input-dependent weight
gu(x)
to the
uth
expert for input
x
, and the final output of the model is a
weighted combination of the outputs of each expert. Hence, each expert contributes differently for
every data point towards the final outcome, which is a key difference from standard ensembles.
Many types of MOE’s exist in the literature [
40
] - the architecture is not standard. For FEAMOE, we
chose this family, with some novel modifications described later, for three main reasons: 1) Suitable
regularization penalties that promote fairness can be readily incorporated into the loss function. 2)
Online learning is possible, so changes in the data can be tracked. Crucially, since localized changes
in data distribution post-deployment may impact only one or a few experts, the other experts may not
need to be adjusted, making the experts localized and only loosely coupled. This allows for handling
drift and avoiding catastrophic forgetting, which is a prime concern in widely used neural network
models [
31
]. 3) Simpler models can be used to fit a more complex problem in the mixture of experts,
as each model needs to fit well in only a limited part of the input space. In particular, even linear
models, which provide very fast SHAP explanations, can be used. The overall mixture of experts,
even with such simple base models (the "experts") often has predictive power that is comparable to a
single complex model such as a neural network, as shown by our experiments as well as in many
previous studies [40].
A motivating toy example of why FEAMOE is needed and how it works is shown in Figure 1.
Consider a linear binary classifier (1a) that has perfect accuracy. The colors represent the ground
truth labels, and green is the positive (desired) class label. The circles are the privileged group and
diamonds are the underprivileged group. As can be seen in the figure, more diamonds receive a
negative outcome and more circles receive a positive outcome. Consider new data that arrives for
predictions. This classifier (1b) not only misclassifies individuals but also gives more underprivileged
individuals that were actually in the positive class a negative outcome, hence inducing bias with
respect to equalized odds. There is drift with respect to accuracy and fairness. A more complex
model (1c) such as a neural network, if retrained, may handle some of these concerns but would be
less explainable.
FEAMOE can address these imperative concerns, as shown in 1d. Trained in an online manner, a
new linear model is added (i.e., an expert) once the new data arrives. The gating network dictates
which region each expert operates in (shown by the blue and pink colors), and FEAMOE is able
to adapt automatically with respect to accuracy and fairness. This dynamic framework enables the
overall model to be fairer, adjust to drift, maintain accuracy, while also remaining explainable since
the decision boundary is locally linear.
2
(a) (b) (c) (d)
Figure 1: A toy example demonstrating the need and use of FEAMOE. The color of every datapoint
corresponds to the original class label. Diamonds represent the underprivileged group and circles
represent the privileged group. (a) Represents a perfectly accurate linear classifier, (b) represents the
same classifier mis-classifying new data points and inducing bias (drift), (c) represents an alternate
non-linear model that corrects for drift but has a complex decision boundary and (d) represents
FEAMOE where the blue and pink regions show the regions of operation for each of the two experts,
separated by the gating network
We show how three fairness constraints–demographic parity, equalized odds, and burden-based
fairness–can be incorporated into the mixture of experts training procedure in order to encourage
fitting fairer models (according to these measures). We use these three popular fairness measures as
illustrative examples to demonstrate the effectiveness of FEAMOE, but our method can be adapted to
incorporate other fairness constraints as well. We then describe a new algorithm for training to account
for drift, where the drift in question can be due to accuracy or fairness. We show experimentally that
by using a set of logistic regression experts, the accuracy of the mixture is comparable to using a
complex model like a neural network. Additionally, we show we can efficiently compute Shapley
value explanations when explanations for every individual expert can be computed quickly. To the
best of our knowledge, this is the first work that addresses the problem of drift with respect to fairness
in a large-scale real world dataset. We then introduce a framework that can flexibly adapt to drifts in
both fairness and accuracy with the added benefit of delivering explanations quickly, while comparing
to the less explainable neural network model class trained in online mode.
The key contributions of this work are: a mixture of experts framework that can incorporate multiple
fairness constraints, a method to handle drift, where drift can be with respect to accuracy or fairness,
empirical evidence of the presence of drift with respect to fairness in a real-world, large-scale dataset,
a theoretical proof that FEAMOE leads to the generation of fast explanations given a suitable choice
of experts, and extensive experimentation on three datasets to show that our method has predictive
performance similar to neural networks while being fairer, handling different types of drift, and
generating faster explanations.
2 Related Work
The mixture of experts (MOE) [
16
,
39
] represent a class of co-operative ensemble models; detailed
surveys on their design and use can be found in [
40
] and [
25
]. Very recently, the deep learning
community has started recognizing and leveraging several advantageous properties that MOE’s have
for efficient design of complex, multi-purpose learners [
30
]. This paper contributes to this expanding
literature by proposing a new algorithm to train this model class to account for both fairness and drift,
and by also adding an explainability module.
Fairness in machine learning is a growing field of research [
14
]. Mitigating biases in models can
be done through pre-processing, in-processing, or post-processing techniques. A description of
these techniques can be found in [
4
]. In-processing techniques for fairness have been gaining
traction [
41
,
26
,
34
]. However, there is limited work on investigating the usefulness of ensemble
models in dealing with biases. [
13
] show that an ensemble of fair classifiers is guaranteed to be
fair for several different measures of fairness, an ensemble of unfair classifiers can still achieve
fair outcomes, and an ensemble of classifiers can achieve better accuracy-fairness trade-offs than a
single classifier. However, they neither provide experimental evidence nor discuss specific methods
to incorporate fairness into ensemble learning. [
24
] develop a method to learn to defer in the case
of unfair predictions. [
5
] use an AdaBoost framework to build a fairer model. [
28
] use adaptive
3
random forest classifiers to account for fairness in online learning, only considering the statistical
parity definition of fairness.
Accounting for drift is a widely explored problem, and is now appearing in commercial products
as well (e.g. model monitoring is a key part of MLOPs) as ML solutions get deployed in business
environments. Details on many such approaches can be found in [
12
,
20
]. Among these approaches,
the one that comes closest to ours is [
37
] which uses a committee of decision trees to account for drift.
However, ensuring fairness in the presence of drift remains an open problem. [
7
] is a very recent
work on achieving a fairer model by building a set of classifiers in the presence of prior distribution
shifts. The method is built for a shift between the training and test distributions, and not for online
learning.
There are many ways to explain a machine learning model [
8
,
27
]. In this paper, we focus on Shapley
values-based explanations, which are widely used in practical applications [
6
]. [
22
] propose the
computation of Shapley values for tree ensembles, which is a faster way to get Shapley values than
through the more broadly applicable method, KernelShap [
21
]. We show that in FEAMOE, the Shap
values for the overall model are just a data-dependent linear combination of the values from individual
experts. Thus the mixture approach does not add any significant complexity to the computation of
feature attribution scores.
3 Theory
We first summarize the original mixture of experts framework and then describe the addition of
fairness constraints. Then, we introduce the algorithm to detect and mitigate data drift when the data
input is sequential (online learning). Thereafter, we show how using the proposed mixture of experts
architecture leads to computing faster Shapley value explanations for the overall non-linear model.
Mixture of Experts (MoE) [
16
] is a technique where multiple experts (learners) can be used to
softly divide the problem space into regions. A gating network decides which expert to weigh most
heavily for each input region. Learning thus consists of the following: 1) learning the parameters of
individual learners and 2) learning the parameters of the gating network. Both the gating network
and every expert have access to the input
x
. The gating network has one output
gi
for every expert
i
. The output vector is the weighted (by the gating network outputs) mean of the expert outputs:
y(x) = Pm
i=1 gi(x)yi(x)
. Consistent with [
16
], the error associated with training the mixture of
experts for case
j
for an accurate prediction is given by:
Ej
acc =log Pigj
ie
1
2||djyj
i||2
,where
yj
i
is the output vector of expert
i
on case
j
,
gj
i
is the proportional contribution of expert
i
to the
combined output vector, and djis the desired output vector.
3.1 Fairness Constraints
In this paper, we incorporate three diverse fairness definitions into the mixture of experts framework:
demographic parity only depends on the model outcome, equalized odds is conditioned on the
ground-truth label, and burden-based fairness depends on the distance of the input to the boundary.
These three popular definitions have been chosen as illustrative metrics; our approach can be readily
extended to several other fairness metrics as well.
For simplicity, we consider a binary classification setting with a binary protected attribute (our
approach readily extends to multi-class and multi-protected attribute problems, where a protected
attribute is a feature such as race or gender). Let
yj
i= 1
be the positive outcome. Let
A= 0
and
A= 1
represent the underprivileged and privileged protected attribute groups, respectively. For a
given dataset
D
, let
Dad
represent all individuals that belong to the protected attribute group
a
and
original class label d.
Statistical parity difference (SPD), which is a measure of demographic parity, measures the difference
between the probability of getting a positive outcome between protected attribute groups [
4
,
35
].
Let
D0
be the set of individuals in the underprivileged group and
D1
be the set of individuals in
the privileged group. Inspired by [
36
], the associated penalty for demographic parity for case
j
is:
Ej
SP D = [jD0](1 Pigiyj
i) + [jD1](Pigiyj
i).
The idea behind this term is that
individuals belonging to the underprivileged group predicted as getting a negative outcome are
4
摘要:

FEAMOE:Fair,ExplainableandAdaptiveMixtureofExpertsShubhamSharmaUniversityofTexasatAustinshubham_sharma@utexas.eduJetteHendersonCognitiveScalejhenderson@cognitivescale.comJoydeepGhoshUniversityofTexasatAustinjghosh@utexas.eduAbstractThreekeypropertiesthataredesiredoftrustworthymachinelearningmodelsde...

展开>> 收起<<
FEAMOE Fair Explainable and Adaptive Mixture of Experts Shubham Sharma.pdf

共20页,预览4页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:20 页 大小:1.04MB 格式:PDF 时间:2025-04-27

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 20
客服
关注