2 N. Rosa et al.
bias with respect to gender) then the learned model will not accurately recognise
and admit women (i.e. unfair behaviour). The second source of bias is due to
some values or demographics of a protected attribute being inherently harder
for AI to recognize than others. For example, it has been shown that even when
training with a balanced dataset, faces with a darker skin tone are harder to
recognize for facial recognition algorithms [30].
Various solutions to the fairness problem have been proposed. We focus on
algorithmic in-processing methods for reducing the bias[9,36,6,31,8,10,5,14]. In-
processing aims to address the bias of a model by applying an extra objective
during training which makes the model bias-aware and consequentially learns a
fairer model. In-processing has proven to be quite effective at reducing the un-
fair behaviour of AI. However, in-processing methods often include extra models
which can increase training cost and complexity [16]; use adversarial training
[9,36,8,6] which has proven to be notoriously unstable [26] or make assump-
tions about the representation space of the model which may not hold in all
cases[10,23,35]. Creating fair AI models is particularly difficult in the computer
vision domain as any problems with extra computational cost and complexity
are exacerbated by the large models utilized. Additionally the high dimensional-
ity of images means they can contain many implicit attributes, which are often
highly correlated to each other and to protected attributes. Disentangling the
implicit factors is extra challenging in these cases.
In this paper, we introduce Bias Accuracy Standard deviation Estimation or
BASE, a novel fairness algorithm, which optimizes a differentiable approxima-
tion of the fairness metric standard deviation of accuracy across demographics
(σacc) to learn an AI model which is fair with respect to equalized odds (EO).
Models that exhibit a low standard deviation of accuracy across demographics
or variance of demographics have the property of equal performance on a target
task regardless of the demographic of the protected attribute. For example, a
facial recognition model which has low variance of demographics for ethnicity,
is equally likely to correctly recognize the identity of a person from an image
regardless of their ethnicity. Reducing the variance of demographics of a model
makes it fairer w.r.t. EO. However, for an AI model that is trained with gradient
based optimization the variance of demographics is difficult to use. This is due to
the accuracy of a single sample - an integral part of the variance of demograph-
ics (Section 2.3) - having an undefined gradient at 0 and being 0 everywhere
else, which leads to zero influence on the model parameters. BASE overcomes
this difficulty by instead using a sigmoid based approximation of accuracy which
we call soft-accuracy inside the variance of demographics metric. This approach
has multiple advantages. Firstly computational efficiency, for example, training
a classifier on images with BASE incurs only the extra computation of calcu-
lating the variance of demographics. Compare this to training a classifier with
knowledge distillation [16] or adversarial debiasing [36], where additional models
are used which incur extra memory usage for the model parameters and gradi-
ents, alongside with extra computation for the forward pass of the additional
model. Secondly, BASE makes no assumptions about the representation space.