2
where the local objective functions fv(x)and the local constraint sets Svare exclusively available to
their corresponding agent. There are multiple motivating applications for this setup. Parameter estimation
and resource allocation in sensor networks [2], [3], [4, chapter 10], fitting models to big datasets [5], [6],
smart grid control [7], and optimal transport [8] are but a few examples.
The standard single-machine optimization algorithms such as projected gradient descent can solve many
instances of the problem in (1). However, in large-scale or privacy-sensitive applications, only distributed
optimization methods are applicable. In a distributed setup, two different architectures can be considered:
centralized, where a central node (agent) coordinates all other agents (workers), and the decentralized
architecture, where there is no central coordinator. To avoid master failure and bottleneck problems,
decentralized architectures have attracted more attention, in the past decade [9]–[12].
In many decentralized optimization schemes, each node voperates on its own realization xvof the
optimization variables. The nodes are to find a common minimizer by communicating their information
through a communication network and computing weighted averages among the neighbors, represented
by so-called gossip matrices [13]. Depending on the nature of communication, the network is represented
by either an undirected or directed graph where the edges denote available communication links between
the agents. In this paper, we consider gossip matrices with directed graphs.
During the past decade, there has been prolific research on decentralized, unconstrained optimization
algorithms [12], [14]. Many well-known algorithms are developed with provable convergence rates under
different assumptions on the objective functions, communication graphs, and the step-size. A state-of-the-
art example is the Push-Pull method [15], [16], utilizing two gossip matrices and a fixed step-size1for
handling directed communication graphs. The optimality gap in Push-Pull is proven to decreases with a
linear rate, when the local objective functions are smooth and strongly-convex. In the absence of strong
convexity and smoothness, no achievable rate of convergence is known for a directed graph.
Compared to the unconstrained setup, the constrained problem, especially with individual constraints as
in (1), is less studied. Several papers consider a simplification by a shared (non-distributed) constraint and
utilize its orthogonal projection operator to find a feasible solution [17]–[19]. Recently, [20] attempted
to address (1) by adapting the Push-Pull algorithm. They demonstrate that the main underlying strategy
(known as gradient tracking) of Push-Pull may not be applied with a fixed step size. However, with a
diminishing step-size they show that the feasibility gap decreases at a sub-linear rate, under the assumption
of smooth and strongly convex functions, but do not discuss the optimality gap. Alternative strategies with
a fixed step size are yet to be discovered. More generally, algorithms with generic convergence properties
in the distributed constrained framework of (1) are still lacking. One main reason is that in the constrained
case, the standard method of analysis based on the Lyapunov (potential) functions becomes complicated. A
decaying step-size can simplify the analysis, but it often results in a dramatic reduction in the convergence
speed [17]. In this paper, we address the above limitations in the distributed convex optimization literature.
Our main contributions are summarized as follows.
A. Contributions
•We propose a novel algorithm, called Double Averaging and Gradient Projection (DAGP), that solves
(1). Our method considers a directed communication graph with individual constraints at each node.
1In general algorithms may use a fixed or a diminishing step size. In practice, fixed step sizes are often superior as diminishing step sizes
require a careful design of a step size schedule, which substantially slows down the underlying algorithm.