Utilizing Explainable AI for improving the
Performance of Neural Networks
Huawei Sun1,3, Lorenzo Servadei1,3, Hao Feng1,3, Michael Stephan1,2, Robert Wille3, Avik Santra1
1Infineon Technologies AG, Neubiberg, Germany
2Friedrich-Alexander-University Erlangen-Nuremberg, Erlangen, Germany
3Technical University of Munich, Munich, Germany
E-mail: {huawei.sun, lorenzo.servadei, avik.santra}@infineon.com
{hao.feng, robert.wille}@tum.de
{michael.stephan}@fau.de
Abstract—Nowadays, deep neural networks are widely used in
a variety of fields that have a direct impact on society. Although
those models typically show outstanding performance, they have
been used for a long time as black boxes. To address this,
Explainable Artificial Intelligence (XAI) has been developing as
a field that aims to improve the transparency of the model and
increase their trustworthiness. We propose a retraining pipeline
that consistently improves the model predictions starting from
XAI and utilizing state-of-the-art techniques. To do that, we use
the XAI results, namely SHapley Additive exPlanations (SHAP)
values, to give specific training weights to the data samples. This
leads to an improved training of the model and, consequently,
better performance. In order to benchmark our method, we
evaluate it on both real-life and public datasets. First, we
perform the method on a radar-based people counting scenario.
Afterward, we test it on the CIFAR-10, a public Computer Vision
dataset. Experiments using the SHAP-based retraining approach
achieve a 4% more accuracy w.r.t. the standard equal weight
retraining for people counting tasks. Moreover, on the CIFAR-10,
our SHAP-based weighting strategy ends up with a 3% accuracy
rate than the training procedure with equal weighted samples.
Index Terms—Radar Sensors, Explainable AI, Deep Learning,
SHapley additive exPlanations
I. INTRODUCTION
Various application areas have been positively affected by
the recent advances of Artificial Intelligence (AI) and Machine
Learning (ML). Among them, fields such as autonomous
driving [1], [2], health tech [3], [4] and robotics [5], [6]
heavily rely on the processing of ML algorithms onto a set
of different sensors. These approaches are typically based
on computationally intensive Deep Learning (DL) strategies,
which involve training millions, or even billions of parameters
to perform a specific task. Although the out-coming results
show high performance, a major problem occurs: As a neural
network gets deeper and deeper, it is also becoming more
complex and thus challenging to be interpreted. To this end,
a neural network is often considered a black box: Even if the
model correctly predicts the given specific input, it is difficult
to explain what causes the correct prediction. This property,
in turn, reduces the trustworthiness of the outcome.
In order to improve a DL system, it is necessary to under-
stand its weaknesses and shortcomings [7]. Approaching this,
XAI focuses on improving the transparency of ML technolo-
gies and increasing their trust. When a model’s predictions
are incorrect, explanatory algorithms can aid in tracing the
underlying reasons and phenomenon. XAI has been researched
for several years, and lots of work has been done in fields
such as Computer Vision (CV) [8], [9], [10] and Natural
Language Processing (NLP) [11], [12]. These algorithms
mainly generate attention maps, which help to highlight the
critical area/words in classifying images or during language
translation. Nevertheless, nowadays DL is widely applied in
less conventional application fields: For example, radar-based
solutions for tasks such as counting people [13], identifying
gestures [14], and tracking [15], as shown in this contribution
[16]. Although the advancements mentioned above success-
fully solve radar-based problems, explaining DL models for
radar signals is still a challenging topic. Additionally, most
XAI algorithms analyze the predictions from a well-trained
model, thus focusing only on the explanatory part. To this end,
a few research contributions move forward by utilizing the
results from XAI for secondary tasks. Layer-wise Relevance
Propagation (LRP), for example, is used for adaptive learning
rate during training in [17], and in [18] the authors prune Deep
Neural Networks (DNNs) and quantize the weights mainly by
Deep Learning Important FeaTures (DeepLIFT). However, to
the best of our knowledge, XAI has not yet been used to
process the dataset and improve the network performance. In
this paper, we first adapt our method to a real-life use case:
Radar-based people counting. Afterward, we show promising
results on the CIFAR-10 dataset [19] to further underlying the
approach’s generality.
Radar-based people counting is a significant application
with high privacy preservation and weather condition indepen-
dence compared to camera-based people counting. However,
the algorithms often underperform the state-of-the-art com-
puter vision methods, and radar data often has the limitation
of low-resolution and room dependency [20], [21]. To this end,
many solutions have been implemented which use DL for this
task in different scenarios [13], [22]. Although performant,
those solutions do not consider how the network obtains the
prediction and which features are essential to explain the
outcome of the task.
This paper introduces a retraining pipeline, which adopts
arXiv:2210.04686v1 [cs.LG] 7 Oct 2022