Probabilistic Verification of Approximate
Algorithms with Unstructured Errors:
Application to Fully Inexact Generalized ADMM
Anis Hamadouche, Yun Wu, Andrew M. Wallace, and Jo˜
ao F. C. Mota
Abstract—We analyse the convergence of an approximate, fully
inexact, ADMM algorithm under additive, deterministic and
probabilistic error models. We consider the generalized ADMM
scheme that is derived from generalized Lagrangian penalty
with additive (smoothing) adaptive-metric quadratic proximal
perturbations. We derive explicit deterministic and probabilistic
convergence upper bounds for the lower-C2nonconvex case as
well as the convex case under the Lipschitz continuity condition.
We also present more practical conditions on the proximal
errors under which convergence of the approximate ADMM to
a suboptimal solution is guaranteed with high probability. We
consider statistically and dynamically-unstructured conditional
mean independent bounded error sequences. We validate our re-
sults using both simulated and practical software and algorithmic
computational perturbations. We apply the proposed algorithm
to a synthetic LASSO and robust regression with k-support
norm regularization problems and test our proposed bounds
under different computational noise levels. Compared to classical
convergence results, the adaptive probabilistic bounds are more
accurate in predicting the distance from the optimal set and
parasitic residual error under different sources of inaccuracies.
Index Terms—Numerical Linear Algebra; Numerical Op-
timization; ADMM; Douglas-Rachford method; Approximate
Computing.
I. INTRODUCTION
The Alternating Direction of Multiplier Method (ADMM)
originally emerged from the early works of Peaceman and
Rachford [2, 1, 3, 4], Glowinski and Marrocco [8], and Gabay
and Mercier [10] on applying numerical procedures to solve
partial differential equations arising in heat conduction and
continuum mechanics problems. Due to its simple implemen-
tation and efficiency in solving large-scale optimization prob-
lems encountered in statistics, and more recent machine learn-
ing problems, ADMM has evolved into different forms and
has been applied to solve more general nonconvex problems
that cannot be solved using conventional methods. ADMM is
also a popular method for online and distributed optimization
[22]. ADMM can be viewed as a decomposition procedure that
exploit the separability of the objective function to reduce the
difficulty of joint minimisation arising from classical methods
of multiplier [5, 6, 7, 9] at the cost of increased number of
iterations [39]. However, this procedure yields efficient local
Work supported by UK’s EPSRC (EP/T026111/1, EP/S000631/1), and the
MOD University Defence Research Collaboration.
Anis Hamadouche, Yun Wu, Andrew M. Wallace, and Jo˜
ao F.
C. Mota are with the School of Engineering & Physical Sci-
ences, Heriot-Watt University, Edinburgh EH14 4AS, UK. (e-mail:
{ah225,y.wu,a.m.wallace,j.mota}@hw.ac.uk).
solutions to subproblems that are coordinated via constraints to
find the global solution [10, 8]. The subproblems are typically
easier to solve than the original problem, and as a matter of
fact, closed-form solutions are often available. For instance,
in the following LASSO problem
minimize
x∈Rn,z∈Rmf(x) := kAx −bk2
2+kzk1,
subject to x−z= 0 ,
(1)
when the data matrix Ahas some circulant structure as those
in signal/image processing, the system can be solved via fast
Fourier transform with insignificant computation cost [21,
26, 30]. However, when the ADMM subproblems do not
possess closed-form solutions, the solution is obtained (or
approximated) iteratively which can lead to increased compu-
tation cost with additional incurred truncation (computational)
errors. This motivates the need for new computationally-aware
variants of the classical ADMM with inexpensive auxiliary
subproblems.
Another feature that makes ADMM technique practically
appealing is its robustness to poorly selected algorithm param-
eters, i.e., the method is guaranteed to converge independently
from the selected parameter. For instance, if the objective
functions are strongly convex and have Lipschitz-continuous
gradients, then the iterates produced by the ADMM algorithm
converge linearly to the optimum in a certain distance metric
no matter which parameter value is selected [39, 32, 41, 38, 40,
43, 46, 62]. For an extensive review of ADMM, the interested
reader is referred to the survey papers by [22, 36, 33].
Before we delve into the details of composite optimization,
ADMM technique and proximal calculus, let us first recall
some basic definitions from the theory of optimization and
proximal calculus.
A. Background
Definition 1 (Convex and Concave functions).A function f:
Rn→R∪ {±∞} is convex if
f(βx +(1−β)y)≤βf (x)+ (1−β)f(y),∀x, y ∈Rn.(2)
with β∈(0,1). If fis convex then −fis concave.
Definition 2 (Proper function).A function is proper of its
value is never −∞ and it is finite somewhere.
arXiv:2210.02094v1 [math.OC] 5 Oct 2022