and Mudag [42] employ multi-step gossip procedures. In order to achieve acceptable
complexity bounds, one may distribute accelerated methods directly [7, 14, 19, 30, 42]
or use a Catalyst framework [21]. Accelerated methods meet the lower complexity
bounds for decentralized optimization [12, 22, 34]. As usual, by complexity we mean
a sufficient number of iterations of the algorithm that guarantee the solution of the
problem with a given accuracy.
Consensus restrictions x1=. . . =xnmay be treated as linear constraints, thus
allowing for a dual reformulation of problem (1). Dual-based methods include dual
ascent and its accelerated variants [34, 38, 40, 44]. Primal-dual approaches like ADMM
[2, 39] are also implementable in decentralized scenarios. In [35], the authors developed
algorithms for non-smooth convex objectives and provided lower complexity bounds
for this case, as well.
Changing topology for time-varying networks requires new approaches to decen-
tralized methods and a more complicated theoretical analysis. The first method with
provable geometric convergence was proposed in [26]. Such primal algorithms as the
Push-Pull Gradient Method [29] and DIGing [26] are robust to network changes and
have theoretical guarantees of convergence over time-varying graphs. Recently, a dual
method for time-varying architectures was introduced in [24].
In [33], it was studied the problem of decentralized optimization with strongly con-
vex smooth objective functions. The authors investigated accelerated deterministic
algorithms under time-varying network constraints with a consensus projection proce-
dure. The distributed stochastic optimization over time-varying graphs with consensus
projection procedure was studied in [32]. The consensus projection procedure made it
possible to use more acceptable parameters of smoothness and strong convexity in the
complexity estimates. The main goal of this paper is to investigate a similar consensus
projection approach for problems with PL-condition. Note that approach [32, 33] is
based on the well-known concept of the inexact (δ, L, µ)-oracle. But in the PL-case,
we will use inexact gradients with additive noise to describe the consensus projection
procedure.
1.2. Our contributions
Summing up, the contribution of this paper is as follows.
•We study the sum-type minimization problem when the objective function sat-
isfies the PL-condition. To solve a such class of problems, we propose a gradient
descent type method with consensus projection procedure (see Algorithm 2) and
access to only inexact gradient. We consider two cases of gradient inexactness:
the bounded deterministic inexactness and the sum of random noise with de-
terministic bias. We estimated the sufficient communication steps and iterations
number to approach the required quality concerning the function and the dis-
tance between the agent’s vectors and their average for both cases.
•We study the decentralized saddle-point problem (with the structure of a sum)
when the objective function satisfies the two-sided PL-condition. For solving a
such generalized class of problems, we proposed a generalization of the MGDA
method (see Algorithm 3) with a consensus procedure. We provided an estima-
tion for the sufficient number of iterations for inner and outer loops to approach
the acceptable quality concerning the function. Also, we estimate the commu-
nication complexity to a distance between the agent’s vectors and their average
for both cases to be small enough. Additionally, we research the influence of
4