
in such imprecisely known parameters (in this case, the mass)
using a nondeterministic framework instead.
Problem statement.
Our goal is to synthesize a controller
that (1) is robust against nondeterminism due to parameter
uncertainty and (2) reasons over probabilities derived from
stochastic noise. In other words, the controller must satisfy a
given specification under any possible outcome of the nonde-
terminism (robustness) and with at least a certain probability
regarding the stochastic noise (reasoning over probabilities).
We wish to synthesize a controller with a probably approxi-
mately correct (PAC)-style guarantee to satisfy a reach-avoid
specification with at least a desired threshold probability.
Thus, we solve the following problem:
Problem.
Given a reach-avoid specification for a stochas-
tic model with uncertain parameters, compute a con-
troller and a lower bound on the probability that, under
any admissible value of the parameters, the specification
is probabilistically satisfied with this lower bound and
with at least a user-specified confidence probability.
We solve this problem via a discrete-state abstraction of the
continuous model. We generate this abstraction by partition-
ing the continuous state space and defining actions that induce
potential transitions between elements of this partition.
Algorithmically, the closest approach to ours is Badings
et al. (2022), which uses abstractions to synthesize controllers
for stochastic models with aleatoric uncertainty of unknown
distribution, but with known parameters. Our setting is more
general, as epistemic uncertainty requires fundamental differ-
ences to the technical approach, as explained below.
Robustness to capture nondeterminism.
The main con-
tribution that allows us to capture nondeterminism, is that
we reason over sets of potential transitions (as shown by
the boxes in Fig. 1), rather than precise transitions, e.g., as
in Badings et al. (2022). Intuitively, for a given action, the
aleatoric uncertainty creates a probability distribution over
sets of possible outcomes. To ensure robustness against epis-
temic uncertainty, we consider all possible outcomes within
these sets. We show that, for our class of models, comput-
ing these sets of all possible outcomes is computationally
tractable. Building upon this reasoning, we provide the fol-
lowing guarantees to solve the above-mentioned problem.
1) PAC guarantees on abstractions.
We show that
both probabilities and nondeterminism can be captured
in the transition probability intervals of so-called interval
Markov decision processes (iMDPs, Givan, Leach, and Dean
2000). We use sampling methods from scenario optimiza-
tion (Campi, Car
`
e, and Garatti 2021) and concentration in-
equalities (Boucheron, Lugosi, and Massart 2013) to compute
PAC bounds on these intervals. With a predefined confidence
probability, the iMDP correctly captures both aleatoric and
epistemic uncertainty.
2) Correct-by-construction control.
For the iMDP, we
compute a robust optimal policy that maximizes the worst-
case probability of satisfying the reach-avoid specification.
The iMDP policy is automatically translated to a controller
for the original, continuous model on the fly. We show that, by
construction, the PAC guarantees on the iMDP carry over to
the satisfaction of the specification by the continuous model.
Contributions.
We develop the first abstraction-based, for-
mal controller synthesis method that simultaneously captures
epistemic and aleatoric uncertainty for models with contin-
uous state and action spaces. We also provide results on the
PAC-correctness of obtained iMDP abstractions and guaran-
tees on the synthesized controllers for a reach-avoid speci-
fication. Our numerical experiments in Sect. 6 confirm that
accounting for epistemic uncertainty yields controllers that
are more robust against deviations in the parameter values.
Related Work
Uncertainty models.
Distinguishing aleatoric from epis-
temic uncertainty is a key challenge towards trustworthy
AI (Thiebes, Lins, and Sunyaev 2021), and has been con-
sidered in reinforcement learning (Charpentier et al. 2022),
Bayesian neural networks (Depeweg et al. 2018; Loquercio,
Seg
`
u, and Scaramuzza 2020), and systems modeling (Smith
2013). Dynamical models with epistemic parameter uncer-
tainty (but without aleatoric uncertainty) are considered
by Yedavalli (2014), and Geromel and Colaneri (2006). Con-
trol of models similar to ours is studied by (Modares 2022),
albeit only for stability specifications.
Model-based approaches.
Abstractions of stochastic mod-
els are a well-studied research area (Abate et al. 2008; Alur
et al. 2000), with applications to stochastic hybrid (Cauchi
et al. 2019; Lavaei et al. 2022), switched (Lahijanian, Ander-
sson, and Belta 2015), and partially observable systems (Bad-
ings et al. 2021; Haesaert et al. 2018). Various tools exist,
e.g., StocHy (Cauchi and Abate 2019), ProbReach (Shmarov
and Zuliani 2015), and SReachTools (Vinod, Gleason, and
Oishi 2019). However, in contrast to the approach of this
paper, none of these papers deals with epistemic uncertainty.
Fan et al. (2022) use optimization for reach-avoid control
of linear but non-stochastic models with bounded distur-
bances. Barrier functions are used for cost minimization in
stochastic optimal control (Pereira et al. 2020). So-called
funnel libraries are used by Majumdar and Tedrake (2017)
for robust feedback motion planning under epistemic uncer-
tainty. Finally, Zikelic et al. (2022) learn policies together
with formal reach-avoid certificates using neural networks
for nonlinear systems with only aleatoric uncertainty.
Data-driven approaches.
Models with (partly) unknown
dynamics express epistemic uncertainty about the underlying
system. Verification of such models based on data has been
done using Bayesian inference (Haesaert, den Hof, and Abate
2017), optimization (Kenanian et al. 2019; Vinod, Israel, and
Topcu 2022), and Gaussian process regression (Jackson et al.
2020). Moreover, Knuth et al. (2021), and Chou, Ozay, and
Berenson (2021) use neural network models for feedback
motion planning for nonlinear deterministic systems with
probabilistic safety and reachability guarantees. Data-driven
abstractions have been developed for monotone (Makdesi,
Girard, and Fribourg 2021) and event-triggered systems (Pe-
ruffo and Mazo 2023). By contrast to our setting, these