To bridge these gaps, we propose a new approach to generate CF explanations for any differentiable
classifier via feasible perturbations. For this, we extend [
8
] by formulating an objective function for
generating CF instances that takes into account two types of feasibility constraints:
•Global feasibilities
: unary and binary monotonic causal constraints extracted from a domain
expert,
•Local feasibilities
: constraints in the form of feature perturbation difficulty values, given
by the end-users.
The objective function is optimized using gradient descent and feasibility constraints are satisfied
during the optimization by rejecting gradient steps that do not satisfy them. It is important to note
that, here we differentiate between end-user and domain expert. An end-user is the individual who is
subject to the decision of the ML model, e.g. a bank customer whose loan application is rejected. A
domain expert, on the other hand, knows the data and the application. We believe domain experts are
naturally able to give feedback on causal relationship among (at least) some features, without being
constrained to know the exact functional relationship.
The same feasibility constraints were also considered in [
7
] for CF generation. They propose a
generative model based on an encoder-decoder framework, where the encoder projects features into a
latent space and the decoder generates CF instances from the latent space. Their approach, however,
requires complete information about the structural causal model including the causal graph and the
structural equations. This assumption is highly restrictive for applicability of the method in real-world
applications. To cope with this issue, [
7
] proposed a data driven approach to approximate unary and
binary monotonic causal constraints and adopt the approximated relationships in the CF generation.
For local feasibility constraints, they considered implicit user preferences, i.e., given a pair of original
instance and CF instance,
(x,x0)
, the user outputs
1
if CF instance is locally feasible and
0
otherwise.
However, since there is no access to the
(x,x0)
query pairs apriori, they approximate the user by
first asking user preferences on some
(x,q)
, where
q
are sample CF instances generated by a CF
generator without considering user preferences, and then learn a model that generates scores for each
pair that mimics user preferences.
Our approach is different from [7] in several aspects:
•
in [
7
], for approximating each binary constraint, the model learns
2
extra parameters. This
hinders the scalability of the method. Furthermore, these approximated binary constraints
could be imprecise as they are learned from the data, while in our approach we rely on
domain experts to provide such constraints which is more reliable,
•
local feasibility constraints are incorporated via implicit feedbacks that are approximated
using a function. These feedbacks are not directly related to the final CF instances to be
generated. This could result in undesirable CF instances that do not satisfy user’s constraints.
On the other hand, we adopt explicit user feedbacks directly into the optimization function,
•
the type of the user feedback considered in [
7
] for local feasibility is difficult to provide
and restrictive. It is difficult to provide since the user must compare the CF instance with
the original instance to find out if perturbations are locally feasible or not. It is restrictive
because the approach provides no tool for the user to state the level of local infeasibilty. As
an example, assume a CF instance is generated by perturbing more than one feature of the
original instance where all but one perturbation satisfy user’s feasibility constraints. In our
approach, user feedbacks are "feature level" and they are not restricted to {0,1},
•
last but not least, [
7
] did not test they approach in a real user study and it is not evident from
the paper how a real user could be adopted in-the-loop to obtain desirable CF explanation.
To explore the effectiveness of our explanations, we design user studies where users are asked to rank
CF instances generated under different conditions. Through these studies, we found that users tend to
give significantly better ranks to CF instances generated by considering global feasibility constraints
compared to the case where such constraints are not considered. Furthermore, CF instances generated
by adopting both local and global feasibility constraints are better than those generated by only
considering global feasibility constraints. However, their difference is not statistically significant.
In summary, we make the following contributions:
2