Stability Via Adversarial Training of Neural Network Stochastic
Control of Mean-Field Type
Julian Barreiro-Gomez Salah Eddine Choutri Boualem Djehiche
Abstract— In this paper, we present an approach to neural
network mean-field-type control and its stochastic stability
analysis by means of adversarial inputs (aka adversarial at-
tacks). This is a class of data-driven mean-field-type control
where the distribution of the variables such as the system
states and control inputs are incorporated into the problem.
Besides, we present a methodology to validate the feasibility of
the approximations of the solutions via neural networks and
evaluate their stability. Moreover, we enhance the stability by
enlarging the training set with adversarial inputs to obtain a
more robust neural network. Finally, a worked-out example
based on the linear-quadratic mean-field type control problem
(LQ-MTC) is presented to illustrate our methodology.
Index Terms— Neural networks, data-driven control, sta-
bility, robustness, supervised machine learning, adversarial
training
I. INTRODUCTION
Mean-field type control is a topic that attracted a lot
of attention since the introduction of mean-field games by
Lasry and Lions in their seminal work [1] and by Caines,
Huang and Malham´
e in [2]. Andersson and Djehiche in
[3] introduced a stochastic mean-field type control problem
in which the state dynamics and the performance criterion
depend on the moments of the state, see also [4] and [5].
Carmona and Delarue [6], and Buckdahn et al. [7], later,
generalized the problem to include the probability law of the
state dynamics. For applications related to mean-field type
control and games problems we cite, among many others,
[8, Chapter 16], [9]–[11] and the references therein.
This class of problems is non-conventional since both the
evolution of the state and often the performance functional
are influenced by terms that are not directly related to the
state or to the control of the decision maker. In a sense,
they model a very large number of agents behaving, all,
similarly to a representative agent. The latter is impacted by
the aggregation of all agents due to the large number. The
aggregation effect is modelled as a mean-field term such as,
among others, the law of the state, the expectation of the
state or its variance.
Solving this problem, analytically is rather challenging
as there are no general analytic methods for this purpose.
Therefore the use of numerical methods is often needed to
Julian Barreiro-Gomez and Salah Eddine Choutri are with NYUAD
Research Institute, New York University Abu Dhabi, PO Box 129188,
Abu Dhabi, United Arab Emirates. (e-mails: jbarreiro@nyu.edu,
choutri@nyu.edu).
Boualem Djehiche works at the Department of Mathematics, KTH,
Stockholm, Sweden. (e-mails: boualem@kth.se).
We gratefully acknowledge support from Tamkeen under the NYU Abu
Dhabi Research Institute grant CG002. We thank Hatem Hajri for the
discussions on adversarial training.
provide approximations to the solutions and several methods
have been suggested for finite horizon mean-field type con-
trol problems (see e.g. [12], [13]). Furthermore, the recent
progress on machine learning technologies made it easier
to test and provide more efficient approximations of the
solutions to complex mean-field type control problems.
The link between deep learning and mean-field type
control and games was recently studied by, among others,
Lauri`
ere, Carmona and Fouque in series of papers (see e.g.
[14]–[17]), where the authors (jointly and/or independently)
proposed algorithms for the solution of mean-field type
optimal control problems based on approximations of the
theoretical solutions by neural networks, using the software
package TensorFlow with its ‘Stochastic Gradient Descent’
optimizer designed for machine learning. However, the sta-
bility of neural networks associated to the mean-field type
control problems was not considered in the literature so far.
In deep learning, there is an increasing interest in studying
and improving the robustness and stability of the trained
neural networks see e.g. [18], where it has been reported
that a simple modification in the input data might fool a
well-trained neural network, returning a wrong output. For
instance, a picture that is previously well-classified by a
trained neural network could be incorrectly classified once
we perturb one or more pixels in it. Such perturbations are
known as adversarial attacks and can help to characterize
how robust and stable a network is. The contribution of
this paper is summarized in three points: training, stability
evaluation, and stability improvement.
Training: we present an indirect and simple method to
train neural networks to learn optimal controls based on data.
Inspired by the work in [15] and [16], we first illustrate how
data can be generated by computationally solving a finite-
time horizon optimal control problem with decision variables
given by the output of the neural network. Then, we design a
data-driven (model-free) mean-field-type control using neural
networks in a supervised learning fashion. The idea is to
design an offline controller (once trained, only a simple
forward run is required) that is more time-efficient than
the conventional online optimization-based control approach,
which can be time consuming depending on the complexity
of the problem. In real life, one can use the data provided by
a traffic application such as google maps to train the neural
network to give optimal paths.
Stability: we borrow the idea of an adversarial attack from
the topic of image classification and draw an analogy in
the context of stochastic dynamical systems. An adversarial
attack, in our sense, is an initial condition that might make
arXiv:2210.00874v1 [math.OC] 27 Sep 2022