16
1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi ? , Student Member, IEEE, Alec Koppel , Member, IEEE, and Ketan Rajawat ? , Member, IEEE Abstract—We consider expected risk minimization in multi- agent systems comprised of distinct subsets of agents operating without a common time-scale. Each individual in the network is charged with minimizing the global objective function, which is an average of the sum of the statistical average loss function of each agent in the network. Since agents are not assumed to observe data from identical distributions, the hypothesis that all agents seek a common action is violated, and thus the hypothesis upon which consensus constraints are formulated is violated. Thus, we consider nonlinear network proximity constraints which incentivize nearby nodes to make decisions which are close to one another but not necessarily coincide. Moreover, agents are not assumed to receive their sequentially arriving observations on a common time index, and thus seek to learn in an asynchronous manner. An asynchronous stochastic variant of the Arrow- Hurwicz saddle point method is proposed to solve this problem which operates by alternating primal stochastic descent steps and Lagrange multiplier updates which penalize the discrepancies between agents. This tool leads to an implementation that allows for each agent to operate asynchronously with local information only and message passing with neighbors. Our main result establishes that the proposed method yields convergence in expec- tation both in terms of the primal sub-optimality and constraint violation to radii of sizes O( T ) and O(T 3/4 ), respectively. Empirical evaluation on an asynchronously operating wireless network that manages user channel interference through an adaptive communications pricing mechanism demonstrates that our theoretical results translate well to practice. I. I NTRODUCTION In emerging technologies such as wireless communications and networks consisting of interconnected consumer devices [2], increased sensing capabilities are leading to new the- oretical challenges to classical parameter estimation. These challenges include the fact that data is persistently arriving in a sequential fashion [3], that it is physically decentralized across an interconnected network, and that the nodes of the network may correspond to disparate classes of objects (such as users and a base station) with different time-scale requirements [4]. In this work, we seek to address this class of problems through extensions of online decentralized convex optimization [5] to the case where the agents of the network may be of multiple different classes, and operate on different time-scales [6]. To address the fact that we seek iterative tools for streaming data, we consider stochastic optimization problems [7], [8]. In this setting, the objective function E[f (x, θ)] is an expectation over a set of functions parameterized by a random variable θ. A. S. Bedi and A. Koppel are with the U.S. Army Research Laboratory, Adelphi, MD, USA. (e-mail: [email protected], akop- [email protected].). K. Rajawat is with the Department of Electrical Engi- neering, Indian Institute of Technology Kanpur, Kanpur 208016, India (e-mail: [email protected]). A part of this work is presented in IEEE CDC 2018 [1]. The objective function encodes, for example, the quality of a statistical parameter estimate. Through a sequence of realiza- tions of a random variable θ t , we seek to find parameters that are good with respect to the average objective. The classical method to address this problem is stochastic gradient descent (SGD), which involves descending along the negative of the stochastic gradient in lieu of the true gradient to circumvent the computation of an infinite complexity expectation [9], [10]. SGD forms the foundation of tools considered in this paper for asynchronous multi-agent settings. Here we seek solutions to stochastic programs in which data is scattered across an interconnected network G =( V ,E ) of agents, each of which is associated with a unique stream of data {θ i t } t0 . Agents i ∈V then seek to find a solution based on local computations only which is as good as one at a centralized location: this setting is mathematically defined by introducing a local copy of a global parameter estimate and then having each agent seek to minimize a global sum of all local objectives i E[f i (x i , θ i )] while satisfying consensus constraints x i = x j for all node pairs (i, j ) ∈E . Techniques that are good for distributed convex optimization, for example, those based on penalty method [11], [12] or on Lagrange duality [13], [14], have in most cases translated into the distributed stochastic domain without major hurdles, as in [12], [15], [16]. In the aforementioned works, consensus constraints are enforced in order to estimate a common decision variable while leveraging parallel processing architectures to achieve computational speedup [17], [18]. Contrariwise, when unique priors on information available at distinct group of agents are available as in sensor [19] or robotic [20] networks, enforcing consensus degrades the statistical accuracy of each agent’s estimate [19]. Specifically, if the observations at each node are independent but not identically distributed, consensus may yield a sub-optimal solution. Motivated by heterogeneous net- worked settings [20], [21] where each node observes a unique local data stream, we focus on the setting of multi-agent stochastic optimization with nonlinear proximity constraints which incentivize nearby nodes to select estimates which are similar but not necessarily equal. In the setting of nonlinear constraints, penalty methods such as distributed gradient descent do not apply [11], and dual or proximal methods require a nonlinear minimization in an inner-loop of the algorithm [14]. Therefore, we adopt a method which hinges on Lagrange duality that avoids costly argmin computations in the algorithm inner-loop, namely primal-dual method [22], also referred to as saddle point method. Alter- native attempts to extend multi-agent optimization techniques

koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

Page 1: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

1

Asynchronous Saddle Point Algorithm forStochastic Optimization in Heterogeneous NetworksAmrit Singh Bedi?, Student Member, IEEE, Alec Koppel†, Member, IEEE, and Ketan Rajawat?, Member, IEEE

Abstract—We consider expected risk minimization in multi-agent systems comprised of distinct subsets of agents operatingwithout a common time-scale. Each individual in the networkis charged with minimizing the global objective function, whichis an average of the sum of the statistical average loss functionof each agent in the network. Since agents are not assumed toobserve data from identical distributions, the hypothesis that allagents seek a common action is violated, and thus the hypothesisupon which consensus constraints are formulated is violated.Thus, we consider nonlinear network proximity constraints whichincentivize nearby nodes to make decisions which are close to oneanother but not necessarily coincide. Moreover, agents are notassumed to receive their sequentially arriving observations on acommon time index, and thus seek to learn in an asynchronousmanner. An asynchronous stochastic variant of the Arrow-Hurwicz saddle point method is proposed to solve this problemwhich operates by alternating primal stochastic descent steps andLagrange multiplier updates which penalize the discrepanciesbetween agents. This tool leads to an implementation that allowsfor each agent to operate asynchronously with local informationonly and message passing with neighbors. Our main resultestablishes that the proposed method yields convergence in expec-tation both in terms of the primal sub-optimality and constraintviolation to radii of sizes O(

√T ) and O(T 3/4), respectively.

Empirical evaluation on an asynchronously operating wirelessnetwork that manages user channel interference through anadaptive communications pricing mechanism demonstrates thatour theoretical results translate well to practice.

I. INTRODUCTION

In emerging technologies such as wireless communicationsand networks consisting of interconnected consumer devices[2], increased sensing capabilities are leading to new the-oretical challenges to classical parameter estimation. Thesechallenges include the fact that data is persistently arriving in asequential fashion [3], that it is physically decentralized acrossan interconnected network, and that the nodes of the networkmay correspond to disparate classes of objects (such as usersand a base station) with different time-scale requirements [4].In this work, we seek to address this class of problems throughextensions of online decentralized convex optimization [5] tothe case where the agents of the network may be of multipledifferent classes, and operate on different time-scales [6].

To address the fact that we seek iterative tools for streamingdata, we consider stochastic optimization problems [7], [8]. Inthis setting, the objective function E[f(x,θ)] is an expectationover a set of functions parameterized by a random variable θ.

A. S. Bedi and A. Koppel are with the U.S. Army ResearchLaboratory, Adelphi, MD, USA. (e-mail: [email protected], [email protected].). K. Rajawat is with the Department of Electrical Engi-neering, Indian Institute of Technology Kanpur, Kanpur 208016, India (e-mail:[email protected]). A part of this work is presented in IEEE CDC 2018 [1].

The objective function encodes, for example, the quality of astatistical parameter estimate. Through a sequence of realiza-tions of a random variable θt, we seek to find parameters thatare good with respect to the average objective. The classicalmethod to address this problem is stochastic gradient descent(SGD), which involves descending along the negative of thestochastic gradient in lieu of the true gradient to circumventthe computation of an infinite complexity expectation [9], [10].SGD forms the foundation of tools considered in this paperfor asynchronous multi-agent settings.

Here we seek solutions to stochastic programs in whichdata is scattered across an interconnected network G=(V,E)of agents, each of which is associated with a unique streamof data {θit}t≥0. Agents i ∈ V then seek to find a solutionbased on local computations only which is as good as one ata centralized location: this setting is mathematically defined byintroducing a local copy of a global parameter estimate andthen having each agent seek to minimize a global sum of alllocal objectives

∑i E[f i(xi,θi)] while satisfying consensus

constraints xi = xj for all node pairs (i, j) ∈ E . Techniquesthat are good for distributed convex optimization, for example,those based on penalty method [11], [12] or on Lagrangeduality [13], [14], have in most cases translated into thedistributed stochastic domain without major hurdles, as in [12],[15], [16].

In the aforementioned works, consensus constraints areenforced in order to estimate a common decision variablewhile leveraging parallel processing architectures to achievecomputational speedup [17], [18]. Contrariwise, when uniquepriors on information available at distinct group of agents areavailable as in sensor [19] or robotic [20] networks, enforcingconsensus degrades the statistical accuracy of each agent’sestimate [19]. Specifically, if the observations at each nodeare independent but not identically distributed, consensus mayyield a sub-optimal solution. Motivated by heterogeneous net-worked settings [20], [21] where each node observes a uniquelocal data stream, we focus on the setting of multi-agentstochastic optimization with nonlinear proximity constraintswhich incentivize nearby nodes to select estimates which aresimilar but not necessarily equal.

In the setting of nonlinear constraints, penalty methods suchas distributed gradient descent do not apply [11], and dualor proximal methods require a nonlinear minimization in aninner-loop of the algorithm [14]. Therefore, we adopt a methodwhich hinges on Lagrange duality that avoids costly argmincomputations in the algorithm inner-loop, namely primal-dualmethod [22], also referred to as saddle point method. Alter-native attempts to extend multi-agent optimization techniques

Page 2: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

2

to heterogeneously correlated problems have been consideredin [23], [24] for special loss functions and correlation models,but the generic problem was online recently solved in [19]with a stochastic variant of the saddle point method.

However, insisting on all agents to operate on a commonclock creates a bottleneck for implementation in practicalsettings because typically nodes may be equipped with differ-ent computational capacity due to power and energy designspecifications, as well as a difference in the sparsity ofeach agent’s data stream. Therefore, we extend the multi-agent stochastic optimization with nonlinear constraints toasynchronous settings. Asynchrony in online optimization hastaken on different forms, such as, for instance, maintaining alocal Poisson clock for each agent [25] or a distribution-freegeneric bounded delay [6], [26], which is our focus.

There’s a significant body of existing work on asynchronousstochastic programming [6], [27], [28]. Much of this workcannot address constrained settings, and many formulationsconflate the distributed setting with parallel computing [29],[30]. Here we focus on the distributed setting. For decen-tralized asynchronous problems, proximal methods such asthe alternating direction method of multipliers (ADMM) havebeen applied [31]–[33]; however, as is well known of ADMM,it’s application to settings with arbitrary convex nonlinearconstraints requires solving a minimization in the inner-loopof the algorithm whose complexity is prohibitive. Approx-imations of proximal methods for nonlinearly constrainedproblems have appeared in the deterministic case [34], [35];however, extending these methods to stochastically constrainedsettings is intricate due to the nonlinear interaction betweennoise and Taylor expansion of the dual. These issues motivateour first-order primal-dual approach, which can address thesetting considered here, namely, the stochastic, asynchronous,decentralized, and heterogeneous setting. To the best of ourknowledge, there is no existing solution to the problem settingconsidered in this work.

Therefore, our main technical contributions are as follows.• In this paper, we extend the primal-dual method of [19],

[22] for multi-agent stochastic optimization problemswith nonlinear network proximity constraints to asyn-chronous settings. The proposed algorithm allows thegradient to be delayed for the primal and dual updatesof the saddle point method.

• The problem formulation in this work is more genericthan [19] in the sense that we allow stochastic constraintsand use of delayed gradients in the algorithm updateswhich is rarely addressed in the literature.

• The main technical contribution of this paper is toprovide mean convergent results for both the globalprimal cost and constraint violation, establishing that theLyapunov stability results of [19] translate successfullyto asynchronous computing architectures that are becom-ing increasingly important in intelligent communicationsystems.

• Furthermore, empirical evaluation on an asynchronouslyoperating cellular network that manages cross-tier inter-ference through an adaptive pricing mechanism demon-strates that our theoretical results translate to practice.

The rest of the paper is organized as follows. A multi-agentoptimization problem without consensus is formulated in Sec-tion II. An asynchronous saddle point algorithm is proposedto solve the problem in Section III. The detailed convergenceanalysis for the proposed algorithm is presented in Section IV.Next, a practical problem of interference management throughpricing is solved in Section V. Section VI concludes the paper.

II. MULTI-AGENT OPTIMIZATION WITHOUT CONSENSUS

The task of asynchronous online decentralized optimiza-tion in heterogeneous networks consists of several technicalchallenges that must be addressed in concert. Namely, theonline aspect of the problem must be addressed with stochasticapproximation; the decentralized component of the problemmust be addressed through use of protocols that requirelocal information and message passing with neighbors only;by heterogeneity, we mean that there may be differencesbetween agents’ local optimization problems that must berespected while incentivizing network coordination; lastly, theasynchrony means that nodes in the network operate in a lock-free manner, processing information as soon as it’s availablerather than synchronizing their computations with all others.We proceed to develop a mathematical framework to clarifyhow these challenges interrelate and may be addressed throughconstrained stochastic programming.

We consider agents i of a symmetric, connected, anddirected network G = (V, E) with |V | = N nodes and|E| = M edges. Each agent is associated with a (non-strongly)convex loss function f i : X × Θi → R that is parameterizedby a p-dimensional decision variable xi ∈ X ⊂ Rp anda random vector θi ∈ Θi ⊂ Rq with unknown probabilitydistributions ρi(θi). The functions f i(xi,θi) for different θi

encodes the merit of a particular linear statistical model xi,for instance, and the random vector θ may be particularizedto a random pair θ = (z,y). In this setting, the random paircorresponds to feature vectors z together with their binarylabels y ∈ {−1, 1} or real values y ∈ R, for the respectiveproblems of classification or regression. Here we address thecase that the local random vector θi represents data which isrevealed to node i sequentially as realizations θit at time t,and agents would like to process this information on the fly.Mathematically this is equivalent to the case where the totalnumber of samples T revealed to agent i is not necessarilyfinite. A possible goal for agent i is the solution of the localexpected risk minimization problem,

xL(i) := argminxi∈X

F i(xi) := argminxi∈Rp

Eθi [f i(xi,θi)] . (1)

where we define F i(xi) := Eθi [f i(xi,θi)] as the localaverage function at node i. We also restrict X to be a compactconvex subset of Rp associated with the p-dimensional param-eter vector of agent i. By stacking the problem (1) across theentire network, we obtain the equivalent problem

xL = argminx∈XN

F (x) := argminx∈XN

N∑i=1

Eθi [f i(xi,θi)] . (2)

where we define the stacked vector x = [x1, . . . ,xN ] ∈XN ⊂ RNp, and the global cost function F (x) :=

Page 3: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

3

∑Ni=1 Eθi [f i(xi,θi)]. We define the global instantaneous cost

similarly: f(x,θ) =∑i f

i(xi,θi).Note that (1) and (2) describe the same problem since the

variables xi at different agents are not coupled to one another.In many situations, the parameter vectors of distinct agents arerelated, and thus there is motivation to couple the estimatesof distinct agents to each other such that one agent maytake advantage of another’s data. Most distributed optimizationworks, for instance, consensus optimization, hypothesize thatall agents seek to learn the common parameters xi for alli ∈ V , i.e., xi = xj , for all j ∈ ni . where ni denotes theneighborhood of agent i. Making all agents variables equalonly makes sense when agents observe information drawnfrom a common distribution, which is the case for industrial-scale machine learning, but is predominantly not the case forsensor [19] and robotic networks [20].

As noted in [19], generally, nearby nodes observe sim-ilar but not identical information, and thus to incentivizecollaboration without enforcing consensus, we introduce aconvex local proximity function with real-valued range of theform hij(xi,xj ,θi,θj) that depends on the observations ofneighboring agents and a tolerance γij ≥ 0. These stochasticconstraints then couple the decisions of agent i to those of itsneighbors j ∈ ni as the solution of the constrained stochasticprogram

x∗ ∈ argminx∈XN

N∑i=1

Eθi[f i(xi,θi)] (3)

s.t. Eθi,θj

[hij(xi,xj ,θi,θj)

]≤ γij , for all j ∈ ni.

Examples of the constraint in the above formulation includeapproximate consensus constraints

∥∥xi − xj∥∥ ≤ γij , quality

of service E[SINR(xi,xj ,θi,θj) ≥ γij ] where SINR isthe signal-to-interference-plus-noise function, relative entropyD(xi || xj) ≤ γij , or budget γmin

ij ≤ xi + xj ≤ γmaxij con-

straints. In this work, we seek decentralized online solutionsto the constrained problem (3) without the assumption thatagents operate on a common time index, motivated by thefact that asynchronous computing settings are common in largedistributed wireless networks. In the next section we turn todeveloping an algorithmic solution that meets these criteria.

Remark 1 The pairwise stochastic constraints in (3) can bereadily generalized to arbitrary neighborhood constraints:

E[hi({xj ,θj}j∈n′

i)]≤ 0 (4)

where the set n′i := ni∪{i} denotes the set of all neighbors ofnode i inlcuding the node i itself. It can be seen that constraintin (3) is a special case of that in (4), with the j-th entry ofthe |ni| × 1 vector function hi(·) defined as[

hi({xj ,θj}j∈n′i)]j

:= hij(xi,xj ,θi,θj)− γij . (5)

More generally, (4) allows the consensus constraints to beimposed on the entire neighborhood of a node. For instancethe approximate version of the consensus constraint xi =(1/|ni|)

∑j∈ni

xj takes the form

∥∥xi − (1/|ni|)∑j∈ni

xj∥∥ ≤ γi. (6)

Such general constraints also arise in communication systemsin form of SINR constraints. For instance, consider a commu-nication system where the interference at node i from j can bewritten as pj(xj , θij) with θij denoting the stochastic channelgain from node j to node i. Then the SINR constraint at nodei is of the form in (4), and is given by

hi({xj , θij}j∈n′i) = γij −

pi(xi, θii)

σ2 +∑j∈ni

pj(xj , θij)(7)

where σ2 denotes the noise power. Hence, the generalizedstochastic problem can be expressed as

x∗ ∈ argminx∈XN

N∑i=1

E[f i(xi,θi)] (8)

s.t. E[hi({xj ,θj}j∈n′

i

)]≤ 0 for all i.

It is mentioned that the convergence analysis is performed forthe generalized problem in (8) for clarity of exposition sincein this more general setting we may vectorize the constraints,while the main results in Section IV are presented for simplerproblem of (3) [see Appendix 0] for increased interpretability.

III. ASYNCHRONOUS SADDLE POINT METHOD

Methods based upon distributed gradient descent andpenalty methods more generally [36]–[38] are inapplicable tosettings with nonlinear constraints, with the exception of [39],which requires attenuating learning rates to attain constraintsatisfaction. On the other hand, the dual methods proposed in[21], [40]–[42] require a nonlinear minimization computationat each algorithm iteration, and thus is impractically costly. Forinstance, it is not possible to solve the primal sub-problemof [21, Eq. (52)] in closed form. Therefore, in this sectionwe develop a computationally light weight method based onprimal-dual method that may operate in decentralized onlineasynchronous settings with constant learning rates that arebetter suited to changing environments.

For a decentralized algorithm, each node i can access theinformation from its neighbors j ∈ ni only. For an online al-gorithm, the stochastic i.i.d quantities of unknown distributionare observed sequentially θit at each time instant t. In additionto these properties, an algorithm is called asynchronous, if pa-rameter updates may be executed with out-of-date informationand the requirement that distinct nodes operate on a commontime-scale is omitted. To develop an algorithm which meetsthese specifications, begin by considering the approximateLagrangian relaxation of (3) stated as

L(x,λ) =

N∑i=1

[E[f i(xi,θi) (9)

+∑j∈ni

λij(hij(xi,xj ,θi,θj

)− γij

)− δε

2(λij)2

]],

where λij is a non-negative Lagrange multiplier associatedwith the non-linear constraint in (3). Here, λ defines thecollection of all dual variables λij into a single vector λ.

Page 4: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

4

Observe that (9) is not the standard Lagrangian of the (3)but instead an regularized Lagrangian due to the presenceof the term −(δε/2)(λij)2. This terms acts like a forgettingfactor on the dual variable with associated parameters δ andε that allows us to control the accumulation of constraintviolation of the algorithm over time, while allowing the dualvariable λij to belong to an unbounded set, which is atypicalof primal-dual methods for nonlinearly constrained problems– see [13], for instance, for a more standard treatment. Thisregularization makes the dual problem strongly concave, whichis often exploited for linearly constrained settings. However,in this work, this strong concavity is not explicitly employed;instead, the regularization allows us to define a custom optimalLagrange multiplier. This observation has appeared in someprior works [19], [43], and proves useful again in this work.See Remark 2 and proofs in the appendices for further details.

The stochastic saddle point algorithm, when applied to(9), operates by alternating primal and dual stochastic gra-dient descent and ascent steps, respectively. We consider thestochastic saddle point method as a template upon which weconstruct an asynchronous protocol. Begin then by definingthe stochastic approximation of the augmented Lagrangianevaluated at observed realizations θit of the random vectorsθi for each i ∈ V:

Lt(x,λ) =

N∑i=1

[f i(xi,θit) (10)

+∑j∈ni

λij(hij(xi,xj ,θit,θ

jt

)−γij

)− δε

2(λij)2

].

The stochastic saddle point method applied to the stochasticLagrangian (10) takes the following form similar to [19] as

xt+1 = PX[xt − ε∇xLt(xt,λt)

], (11)

λt+1 =[λt + ε∇λLt(xt,λt)

]+, (12)

where ∇xL(xt,λt) and ∇λL(xt,λt), are the primal anddual stochastic gradients1 of the augmented Lagrangian withrespect to x and λ, respectively. These are not the actualgradients of (9) rather are stochastic gradients calculated atthe current realization of the random vectors θit for all i. Thecomponent wise projection for a vector x on to the givencompact set X is here denoted by PX (x). Similarity, [·]+represents the component wise projection on to the positiveorthant RM+ . An important point here is that the methodstated in (11) - (12) can be implemented with decentralizedcomputations across the network, as stated in [19, Proposition1]. Here ε > 0 is a constant positive step-size.

Observe that the implementation of the [19, Algorithm 1],which is defined by (11)-(12), it is mandatory to perform theprimal and dual updates at each node with a common timeindex t. The update of primal variable at node i requires thecurrent gradient of its local objective function ∇xif i(xit,θ

it)

and current gradient from all the neighbors j ∈ ni of node ias ∇xihij

(xit,x

jt ,θ

it,θ

jt

). This availability of the gradients

1Note that these may be subgradients if the objective/ constraint functionsare non-differentiable. The proof is extendable to non-differentiable cases.

Nbhd.

Node

xj0

xj[1]j

xj[2]j

xi0 xi

1 xi2

Grad0 Grad[1] Grad[2]

λji0

λij0

λji1

λij1

λji2

λij2

xi0

xj0

xi[1]i

xj[1]j

xi[2]i

xj[2]j

Fig. 1: Message passing in proposed algorithm. The red blockrepresents the gradients required as mentioned in step 4 of theAlgorithm 1 which depends upon the realizations θi

[t]iand θj

[t]jat

time step t.

Algorithm 1 ASSP: Asynchronous Stochastic Saddle PointRequire: initialization x0 and λ0 = 0, step-size ε, regularizer δ

1: for t = 1, 2, . . . , T do2: loop in parallel agent i ∈ V3: Send dual vars. λij

t to nbhd. j ∈ ni

4: Obtain delayed gradients ∇xif i(xi[t]i,θi

[t]i),

∇xihij(xi[t]i,xj

[t]j,θi

[t]i,θj

[t]j

)and constraint function

hij(xi[t]i,xj

[t]j,θi

[t]i,θj

[t]j

).

5: Update xit+1 using (13) local parameter xi

t

xit+1 =PX

[xit − ε

(∇xif

i(xi[t]i ,θ

i[t]i)

+∑j∈ni

(λijt + λji

t

)∇xih

ij(xi[t]i ,x

j[t]j,θi

[t]i ,θj[t]j

))]6: Update dual variables at each agent i [cf. (14)]

λijt+1=

[(1− ε2δ)λij

t + ε(hij(xi[t]i ,x

j[t]j,θi

[t]i ,θj[t]j

))]+.

7: end loop8: end for

from the neighbors on a common time-scale is a strong as-sumption that insists upon perfect communications, similarityof computational capability of distinct nodes, and similarlevels of sparsity among agents’ data that are violated in largeheterogeneous systems. This limitation of synchronized meth-ods motivates the subsequent development of asynchronousdecentralized variants of (11)-(12)

In particular, to ameliorate the computational bottleneckassociated with synchronized computation and communicationrounds among the nodes, we consider situations in which ob-servations and updates are subject to stochastic delays, i.e., anasynchronous processing architecture. These delays take theform of random delays on the gradients which are used for thealgorithm updates. We associate to each node i in the networka time-dependent delay τi(t) for its stochastic gradient. Sincethe gradient corresponding to node i are delayed by τi(t), itimplies that the received gradient corresponds to t−τi(t) timeslot which we denote as [t]i. Rather than waiting for the current

Page 5: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

5

gradient at time t, agent i instead uses the delayed gradientfrom the neighboring nodes at time [t]j for its update at time[t]i. This leads to the following asynchronous primal updatefor stochastic online saddle point algorithm at each node i

xit+1 =PX[xit − ε

(∇xif i(xi[t]i ,θ

i[t]i) (13)

+∑j∈ni

(λijt + λjit

)∇xihij

(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

))].

Likewise, the dual update for each edge (i, j) ∈ E is

λijt+1=[(1− ε2δ)λijt + ε

(hij(xi[t]i ,x

j[t]j,θi[t]i,θ

j[t]j

))]+. (14)

Note that to perform the asynchronous primal updates atnode i in (13), delayed primal gradients ∇xif i(xi[t]i ,θ

i[t]i)

and∇xihij(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

)are utilized. Similarity, the

dual delayed gradient is utilized for the update in (14). For theconsistency in the algorithm implementation, it is assumedthat at each node i, only the recent received copy of thegradient is kept and used for the update. Equivalently, thiscondition can be mentioned as [t]i ≥ [t−1]i which implies thatτi(t) ≤ τi(t−1)+1. For brevity, we will use the notation [t] asa collective notation for all the delayed time instances as [t] :=[[t]1; · · · ; [t]N ]. The asynchronous algorithm is summarized inAlgorithm 1. The practical implementation of the proposedasynchronous algorithm is explained with the help of diagramin Fig.1. As described in figure, each node receives delayedparameters, gradients and carries out the updated accordingly.The convergence guarantees for the proposed algorithm areshown to hold as t − [t]i ≤ τ is finite (assumption 6) forall i and t. Further details of how the algorithm works atthe link layer of the network are summarized in Algorithm 2.For the algorithm execution, it is assumed that the form ofthe function f i and hij is known at node i. Therefore, thecalculation of the gradient is being performed at node i. Tocalculate the gradient at node i for time instant t, it is requiredto estimate the local random quantity θit and receive xjt andθjt from all of its neighbor nodes. This estimate of the randomvariable and communication of primal variables from neighbornodes results in a bottleneck for the execution of synchronousversion of the proposed algorithm. We relax this condition andallows the use of previously calculated gradient at previouslyavailable variables. The algorithm execution steps to be takenat each node i at time instant t are summarized here inAlgorithm 1 presented in this response sheet. From Algorithm1, note that the steps 2 and 3 are optional, which means thatit is not required to carry out them at each time iterationresulting in asynchrony in the network. But they should happeninfinitely often to keep the overall delay bounded as stated inAssumption 6. The convergence results presented in [19] canbe obtained as a special case with τi(t) = 0 from the resultsdeveloped in this paper. This shows the generalization of theexisting results in literature.

Before proceeding with the convergence analysis, in order tohave a tractable derivation, let us define the compact notation

Algorithm 2 Operation at node i at time instant tRequire: initialization x0 and λ0 = 0, step-size ε, regularizer δ

1: for t = 1, 2, . . . , T do2: Send dual variable λij

t to neighborhood nodes j ∈ ni andreceive λji

t

3: (Optional) Receive (xjt ,θ

jt) from the neighbors j ∈ ni

4: (Optional) Calculate gradients ∇xif i(xit,θ

it),

∇xihij(xit,x

jt ,θ

it,θ

jt

)and constraint function

hij(xit,x

jt ,θ

it,θ

jt

)5: Update xi

t+1 using local parameter xit

xit+1 = PX

[xit − ε

(∇xif

i(xi[t]i ,θ

i[t]i) +

∑j∈ni

(λijt

+ λjit

)∇xih

ij(xi[t]i ,x

j[t]j,θi

[t]i ,θj[t]j

))]6: Update dual variables at each node i for all j ∈ ni

λijt+1=

[(1− ε2δ)λij

t + ε(hij(xi[t]i ,x

j[t]j,θi

[t]i ,θj[t]j

))]+.

7: end for

for the primal and dual delayed gradient as follows

∇xiLi[t](xi[t]i ,x

j[t]j,λit,λ

jt

):=(∇xif i(xi[t]i ,θ

i[t]i) (15)

+∑j∈ni

(λijt + λjit )∇xihij(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

))which follows from (13) and

∇λiLi[t](xi[t]i ,x

j[t]j,λit

):=hij

(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

)−ε2δλijt

(16)

follows from (14). These definitions of (15) and (16) will beused in rest of the places in this paper.

IV. CONVERGENCE IN EXPECTATION

In this section, we establish convergence in expectation ofthe proposed asynchronous technique in (13)-(14) to a primal-dual optimal pair of the problem formulated in (3) whenconstant step sizes are used. Specifically, a sublinear bound onthe average objective function optimality gap F (xt)− F (x∗)and the network-aggregate delayed constraint violation isestablished, both on average. The optimal feasible vector x∗

is defined by (3). It is shown that the time-average primalobjective function F (xt) converges to the optimal value F (x?)at a rate of O(1/

√T ). Similarly, the time-average aggregated

delayed constraint violation over network vanishes with theorder of O(T−1/4), both in expectation, where T is thefinal iteration index. To prove convergence of the stochasticasynchronous saddle point method, some assumptions relatedto the system model and parameters are required which westate as follows.

Assumption 1 (Network connectivity) The network G is sym-metric and connected with diameter D.

Assumption 2 (Existence of Optima) The set of primal-dualoptimal pairs X ∗ × Λ∗ of the constrained problem (3) hasnon-empty intersection with the feasible domain XN × RM+ .

Page 6: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

6

Assumption 3 (Stochastic Gradient Variance) The instanta-neous objective and constraints for all i and t satisfy

E∥∥∇xif i(xi,θit)

∥∥2 ≤ σ2f (17)

E∥∥∥∇xihij

(xi,xj ,θi[t]i ,θ

j[t]j

)∥∥∥2 ≤ σ2h (18)

which states that the second moment of the norm of objectiveand constraint function gradients are bounded above.

Assumption 4 (Constraint Function Variance) For the instan-taneous constraint function for all pairs (i, j) ∈ E and t overthe compact set X , it holds that

max(xi,xj)∈X

E[(hij(xi,xj ,θit,θ

jt

)2)]≤ σ2

λ (19)

which implies that the maximum value the constraint functioncan take is bounded by some finite scalar σ2

λ in expectation.

Assumption 5 (Lipschitz continuity) The expected objectivefunction defined in (2) satisfies

‖F (x)− F (y)‖ ≤ Lf ‖x− y‖ . (20)

for any (x,y) ∈ RNp.

Assumption 6 (Bounded Delay) The delay τi(t) associatedwith each node i is upper bounded: τi(t) ≤ τ for some τ <∞.

Assumption 1 ensures that the graph is connected and therate at which information diffuses across the network is finite.This condition is standard in distributed algorithms [37], [40].Moreover, Assumption 2 is a Slater’s condition which makessure the existence of an optimal primal-dual pair within thefeasible sets onto which projections occur which are necessaryfor various quantities to be bounded. It has appeared in var-ious forms to guarantee existence of solutions in constrainedsettings [44]. Assumption 3 assumes an upper bound on themean norm of the primal and dual stochastic gradients, whichis crucial to developing the gradient bounds for the Lagrangianused in the proof. Assumption 4 yields an upper bound onthe maximum possible value of the constraint function inexpectation similar to that of [43], and is guaranteed to holdwhen X is compact and hij is Lipschitz. Assumption 5 isrelated to the Lipschitz continuity of the primal objectivefunction. Assumption 6 ensures that the delay is alwaysbounded by τ , which holds in most wireless communicationsproblems and autonomous multi-agent networks [45].

For the analysis to follow, we first derive bounds on themean square-norms of the stochastic gradients of the La-grangian. Thus, consider the mean square-norm of the primalstochastic gradient of the Lagrangian, stated as:

E[‖∇xLt(x,λ)‖2]≤2N[maxi

E∥∥∇xif i(xi,θit)

∥∥2] (21)

+ 2M2‖λ‖2 max(i,j)∈E

E∥∥∥∇xihij

(xi,xj ,θit,θ

jt

)∥∥∥2Now apply Assumptions in 3 to the mean square-norm termsin (21) to obtain

E[‖∇xLt(x,λ)‖2] ≤ 2Nσ2f + 2M2 ‖λ‖2 σ2

h

≤ 2(N +M2)L2(1 + ‖λ‖2) (22)

where, L2 = max(σ2f , σ

2h). Similarly for the gradient with

respect to the dual variable λ, we have

E[‖∇λLt(x,λ)‖2

]≤2Mmax

(i,j)∈EE[(hij(xi,xj ,θit,θ

jt ))

2]+2δ2ε2‖λ‖2

≤2Mσ2λ +2δ2ε2‖λ‖2 (23)

The bounds developed in (22) and (23) are in terms of thenorm of the dual variable and utilizes the Assumptions 3 and4. It is important to note that these bounds are for arbitrary xand λ, therefore holds for any realization of the primal xt anddual variables λt. Before proceeding towards the main lemmasand theorem of this work, a remark on the importance of thebounds in (22) and (23) is due.

Remark 2 Conventionally, in the analysis of primal-dualmethods, the primal and dual gradients are bounded by con-stants, due to, for instance, projection onto compact primaland dual domains. See [13] as an example of this line ofreasoning. In contrast, here our upper-estimates depend uponthe magnitude of the dual variable λ which allows us to avoiddual set projections onto a compact set, and instead operatewith unbounded dual sets RM+ . We are able to do so viaexploitation of the dual regularization term −(δε/2)‖λ‖2 thatallows us to control the growth of the constraint violation.

Subsequently, we establish a lemma for the instantaneous La-grangian difference L[t](x[t],λ)− L[t](x,λt) by a telescopicquantity involving the primal and dual iterates, as well as themagnitude of the primal and dual gradients. This lemma iscrucial to the proof of our main result at the end of this section.

Lemma 1 Under the Assumptions 1 - 6 , the sequence(xt,λt) generated by the proposed asynchronous stochasticsaddle point algorithm in (13)-(14) is such that for a constantstep size ε, the instantaneous Lagrangian difference sequenceL[t](x[t],λ)− L[t](x,λt) satisfies the decrement property

L[t](x[t],λ)− L[t](x,λt)

≤ 1

(‖xt−x‖2−‖xt+1−x‖2 + ‖λt − λ‖2−‖λt+1−λ‖2

)+ε

2

(∥∥∥∇λL[t](x[t],λt)∥∥∥2 + ‖∇xL[t](x[t],λt)‖2

)+ 〈∇xL[t](x[t],λt), (x[t] − xt)〉 (24)

Proof: See Appendix A. �

Lemma 1 exploits the fact that the stochastic augmentedLagrangian is convex-concave with respect to its primal anddual variables to obtain an upper bound for the differenceL[t](x[t],λ) − L[t](x,λt) in terms of the difference betweenthe current and the next primal and dual iterates to a fixedprimal-dual pair (x,λ), as well as the square magnitudesof the primal and dual gradients. Observe that here, relativeto [19][Proposition 1], an additional term is present whichappears that represents the directional error caused by asyn-chronous updates of Algorithm 1. This contractive property isthe basis for establishing the convergence of the primal iteratesto their constrained optimum given by (3) in terms of mean

Page 7: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

7

SC 3MU 2

Macro Base Station

(MBS)

SC 1

MU 1

SCUSC 2

Fig. 2: Heterogeneous cellular network with one MBS, two MUs, andthree SCBSs with each serving one, two, and one SCU, respectively.

objective function evaluation and mean constraint violationwith constant step-size selection. Before proceeding towardsour main theorem, we establish an additional lemma whichsimplifies its proof and clarifies ideas.

Lemma 2 Denote as (xt,λt) the sequence generated by theasynchronous saddle point algorithm in (13)-(14) with stepsizeε. If Assumptions 1 - 6 holds, then it holds that

E[ T∑t=1

[F (xt)−F (x)]+∑

(i,j)∈E

λij

(T∑t=1

hij(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

))

−(δεT

2+

1

)(λij)2

]≤ 1

2ε‖x1−x‖2 +

εTK

2(25)

where the constant K is defined in terms of system parametersas

K := Mσ2λ + (N +M)L[(1/2)L+ τ2(2L+ Lf )]. (26)

Proof: See Appendix B. �

Lemma 2 is derived by considering Lemma 1, comput-ing expectations, and applying (22) and (23). This lemmadescribes the global behavior of the augmented Lagrangianwhen following Algorithm 1, and may be used to establishconvergence in terms of primal objective optimality gap andaggregated network constraint violation, as we state in thefollowing theorem.

Theorem 1 Under the Assumptions 1 - 6 , denote (xt,λt) asthe sequence of primal-dual variables generated by Algorithm1 [cf. (13)-(14)]. When the algorithm is run for T total itera-tions with constant step size ε = 1/

√T , the average time ag-

gregation of the sub-optimality sequence E [F (xt)− F (x∗)],with x∗ defined by (3), grows sublinearly with T as

T∑t=1

E[F (xt)−F (x∗)]≤ O(√T ). (27)

Likewise, the delayed time aggregation of the average con-strain violation also grows sublinearly in T as∑

(i,j)∈E

E

[[ T∑t=1

(hij(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

)− γij) ]

+

]≤ O(T 3/4).

(28)

Proof: See Appendix C. �

Theorem 1 presents the behavior of the Algorithm 1 whenrun for T total iterations with a constant step size. Specifically,the average aggregated objective function error sequence is up-per bounded by a constant time

√T sequence. This establishes

that the expected value of the objective function E [F (xt)]will become closer to the optimal F (x?) for larger T . Similarbehavior is shown by the average delayed aggregated networkconstraint violation term. The key innovation establishingthis result is the bound on the directional error caused byasynchrony. We achieve this by bounding it in terms of thegradient norms and dual variable λ and exploiting the factthat the delay is at-worst bounded.

These results are similar to those for the unconstrainedconvex optimization problems with sub-gradient descent ap-proach and constant step size. For most of the algorithms inthis context [46, Section 2.2, eqn. 2.19], or [47, Section 4],convergence to the neighborhood of size O(εT ) is well known.For such algorithms, the primal sub-optimality is shown ofthe order O(εT ) is shown and the radius of suboptimalityis optimally controlled by selecting ε = 1/

√T [48]. The

bound O(T 3/4) on the constraint violation aggregation iscomparable to existing results for synchronized multi-agentonline learning [43] and stochastic approximation [19]. Themain motivation to present our results in this manner is toemphasize the conceptual link between mean convergence ofstochastic algorithms with regret analysis of online methods[48], although they are mathematically distinct. To transformthe later into the former, one requires the convex costs ofNature to be parameterized by realizations of independentand identically distributed random variables of a commonstationary distribution. It remains an open challenge to han-dle asynchrony in this online setting, as the distribution-free framework requires more stringent assumptions on thebehavior of the time-delay and the degradation it inflicts onthe algorithm sub-optimality.

It is remarked that the size of the network and magnitude ofthe delay influences the learning constants [c.f. (66)]. There isa linear dependence on the number of nodes, a quadratic de-pendence on the number of edges, and a linear dependence onthe maximum delay. While these are pessimistic estimates, weobserve reasonable performance experimentally as provided inSec.V. Moreover, it is possible to extend the result in (27)to show that the convergence results of the average objectivefunction error sequence also holds for the running averageof primal iterates xT := 1

T

∑Tt=1 xt as in [19, Corollary 1].

However, obtaining such a result for the constraint violationin (28) is not straightforward due to the presence of delayedprimal variables. Subsequently, we turn to studying the em-pirical performance of Algorithm 1 for developing intelligentinterference management in communication systems.

Page 8: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

8

V. INTERFERENCE MANAGEMENT THROUGH PRICING

The rising number of cellular users has fueled the increasein infrastructure spending by cellular operators towards betterserving densely populated areas. In order to circumvent thenear-absolute limits on spectrum availability, the current andfuture generations rely heavily on frequency reuse via smallcells and associated interference management techniques [49]–[51]. This work builds upon the pricing-based interferencemanagement framework proposed in [51]. We consider hetero-geneous networks with multiple autonomous small cell users.Under heavy load situations, the macro base station (MBS)may assign the same operating frequency to multiple butgeographically disparate small cell base stations (SCBS) andmacro cell users (MU). The base station regulates the resultingcross-tier interference (from SCBS to MU) by penalizingthe received interference power at the MUs. Consequently,the SCBSs coordinate among themselves and employ powercontrol to limit their interference at the MUs. This sectionconsiders the pricing problem from the perspective of the BSthat seeks to maximize its revenue.

A. Problem formulation

Consider the network depicted in Fig. 2, consisting of aMBS serving M MU users and N SCBSs [51]. Each MU isassigned a unique subchannel, indexed by i ∈ {1, . . . ,M}.At times of high traffic, the BS also allows the SCBSs touse these M subchannels, so that the n-th SCBS may serveKn ≤M SCUs. In other words, at each time slot, a particularsubchannel is used by MU i and a non-empty set of SCBSsNi ⊂ {1, . . . , N}. Denoting the channel gains between then-th SCBS and i-th MU by gni, it follows that the totalinterference at the MU i is given by Ii :=

∑n∈Ni

gnipin where

pin is the transmit power of SCBS n while using subchannelassigned to MU i. The BS regulates this cross-tier interferenceby imposing a penalty xin on the SCBSs n ∈ Ni. The totalrevenue generated by the BS is therefore given by

M∑i=1

∑n∈Ni

xingnipin (29)

which the BS seeks to maximize. The BS also adheres tothe constraint that the total penalty imposed on each SCBS iswithin certain limit, i.e.,

Cmin ≤∑

i:n∈Ni

xin ≤ Cmax. (30)

The limit on the maximum and minimum penalties can also beviewed as a means for BS to be fair to all small cell operators.The power allocation at the SCBSs is governed by their localtransmission costs, denoted by c per unit transmit power, andthe interference prices levied by the BS. As in [51], each SCBSsolves a penalized rate minimization subproblem, resulting inthe power allocation

pin =((W/(cµn + νnx

in))− (1/hin)

)+

(31)

for all n ∈ Ni and 1 ≤ i ≤ M . Here, hin is the channelgain between n-th SCBS and its scheduled user, µn and

νn represent SCBS-specific parameters used to trade-off theachieved sum rate against the transmission costs and W is thebandwidth per subcarrier. Finally, the channel gains gni andhin are not known in advance, so the BS seeks to solve thefollowing stochastic optimization problem:

max{xi

n}

M∑i=1

∑n∈Ni

E[xingnip

in(xin, h

in)]

(32a)

s. t.∑n∈Ni

E[gnip

in(xin, h

in)]≤ γi 1 ≤ i ≤M (32b)

Cmin ≤∑

i:n∈Ni

xin ≤ Cmax 1 ≤ n ≤ N (32c)

Here, γi is the interference power margin and observe that theinterference constraint is required to hold only on an average,while the limits on the interference penalties are imposed atevery time slot. It is remarked that the similar pricing basedinterference management scheme is considered is [51] but withthe assumption that the distribution of the random variables isknown. For this work, we omit such assumption and proposean online solution to stochastic optimization problem in (32).

Algorithm 3 Online interference management through pricingRequire: initialization x0 and λ0 = 0, step-size ε, regularizer δ

1: for t = 1, 2, . . . , T do2: loop in parallel for all MU and SCBS user3: Send dual vars. λm

t to nbhd.4: Observe the delayed primal and dual (sub)-gradients5: Update the price xin,t+1 at SCBS n as

xin,t+1 =PXn

[xin,t+ ε

(gni,[t]

[W (cµn+νnλ

it)

(cµn+νnxin,[t])2 −

1

hin,[t]

]·1(xin,[t])

)]6: Update dual variables at each MU i [cf. (14)]

λit+1 =

(1−δε2)λit−ε

γi−∑n∈Ni

gni,[t]

(W

cµn + νnxin,[t]

− 1

hin,[t]

)+

+

.

7: end loop8: end for

B. Solution using stochastic saddle point algorithm

It can be seen that the stochastic optimization problemformulated in (32) is of the form required in (3) with Xcapturing the constraint in (32c). Since the random variableshin and gni,t have bounded moments, the assumptions inSection IV can be readily verified. Further, the saddle pointmethod may be applied for solving (32). To do so, we use thepreceding definition of the power function pin defined as in(31), and associating dual variable λi with the i-th constraintin (32), the stochastic augmented Lagrangian is given by

Lt(x,λ) =

M∑i=1

∑n∈Ni

xingni,tpin(xin, h

in,t) (33)

+

M∑i=1

λi[γi−∑n∈Ni

gni,tpin(xin, h

in,t)]+

δε

2‖λ‖2

Page 9: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

9

Iterations, t0 200 400 600 800 1000

Optimality

gap/t

101

102

Asynchronous

Synchronous

(a) Average objective sub-optimality vs. iteration t

Iterations, t0 200 400 600 800 1000

Constraintviolation/t

20

30

40

50 Asynchronous

Synchronous

(b) Average constraint violation vs. iteration t

Fig. 3: Algorithm 3 applied to a 5G cellular network with two MBS user and three SCBSs. The y axis of first figure is1t

∑tu=1 E[F (xu)−F (x∗)] and of second figure is (1/t)E

[[∑tu=1

(hi({xj ,θj

t}j∈n′i

))]+

]. Observe that both the asynchronous and

synchronous implementations attain convergence. Thus, we may solve decentralized online learning problems without a synchronized clock.

where x collects the variables {xin}Mi=1,n∈niand λ collects

the dual variables {λi}Mi=1.The asynchronous saddle point method for pricing-based

interference management in wireless systems then takes theform of Algorithm 3 with the modified projection defined as

PXn(u) := min

y‖y − u‖

s. t. Cmin ≤ 〈1,y〉 ≤ Cmax. (34)

In order to get the primal and dual updates of step 5 and 6,note that the subgradient of the Lagrangian in (33) with respectto primal variables is given by

∂xinLt(xin, λi) := gni,t

[W (cµn + νnλ

i)

(cµn + νnxin)2 −

1

hin,t

]· 1(xin)

and the gradient of the Lagrangian with respect to the dualvariable λi is given by

∇λiLt(xin, λi) =γi−∑n∈Ni

gni,t

(W

cµn + νnxin,t− 1

hin,t

)+

+ δελi.

Observe here that the implementation in Algorithm 3 allowsthe primal and dual updates to be carried out in a decentralizedmanner at each SCBS. On the other hand, the MU carries outthe dual updates. Consequently both, the SCBSs and the BSsmay utilize old price iterates xin,[t]. Therefore, the proposedalgorithm is decentralized in terms of execution at each MU.

For the simulation purposes, we consider a cellular networkwith M = 10 MBSs and N = 11 SCBSs with indices{m1;m2; · · · ;m10} and {s1; s2; · · · ; s11}, respectively. Thescenario considered is similar to as shown in Fig. 2, that isnodes {s1, s2} are in the neighborhood of node m1, {s2, s3}are in the neighborhood of node m2 and so on. The randomchannel gains gni and hin are assumed to be exponentiallydistributed with mean µ = 1 and 3, respectively. The minimumand maximum values Cmin = 0.1 and Cmax = 20. The otherparameter values are W = 5MHz, γi = −4 dB, δ = 10−4,c = 0.1, µn = νn = 1, and ε = 0.01. The maximum delayparameter is τ = 10.

Fig. 3a shows the difference of running average of primalobjective from its optimal value. It is important to note that

Iteration, t0 500 1000 1500

Avg

. rev

enue

30

40

50

60

70

80γi= 2 dB

γi= 0 dB

Fig. 4: Average revenue for different interference power margin.

the difference goes to zero as t→∞. The result for both syn-chronous and asynchronous algorithm algorithms are plotted.The optimal value to plot Fig. 3a is obtained by running thesynchronous algorithm for long duration of time and utilizingthe converged value as the optimal one. We observe thatrunning the saddle point method without synchrony breaksthe bottleneck associated with heterogeneous computing ca-pabilities of different nodes, although it attains slightly slowerlearning than its synchronized counterpart. Fig. 3b shows thebehavior of constraint violation term derived in (28) for arandomly chosen MBS. In the sample path of the empiricalaverage constraint violation, the trend of sublinear growth ofobjective sub-optimality from Fig. 3a is further substantiatedby convergence in expectation of the constraint violation as theiteration index t increases. We observe that the performancereduction of asynchronous operations relative to synchronousones is smaller with respect to constraint violation as comparedwith primal-suboptimality, corroborating the rate analysis ofTheorem 1.

Fig. 4 shows that the average value of the revenue generatedby MBS converges to a higher value for higher interferencepower margin γi = 4 dB. It shows the advantage of proposedinterference management scheme which exploits the higherallowed interference as a resource to generate revenue for theMBS. To further demonstrate the advantage of asynchronous

Page 10: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

10

Iterations, t0 100 200 300 400 500

Tot

al r

even

ue

101

102

103

104

Pipelined implementation (asyn.)Serial implementation (sync.)

Fig. 5: Total revenue generated for synchronous and asynchronousalgorithms.

implementation over the synchronous implementation, wecompare the total revenue of both the schemes in Fig.. 5.Note that the synchronous implementation is akin to serialimplementation while asynchronous implementation allowspipelined execution of the updates. We consider a scenario inwhich each node introduces a processing delay of 7 clock cy-cles (equivalently τ = 7) in the network. For the synchronousalgorithm, each node must wait for 7 clock cycles to get therequired information to perform the primal and dual updates.For instance, over 1000 time slots, due to the waiting time, thesynchronous updates could only be performed 152 times as thenodes are waiting at other times. On the other hand, a pipelinedimplementation is possible with the asynchronous algorithmsince at each time t, the node may use information from (t−7)-th iteration. Consequently, the asynchronous updates occur atevery time slot, resulting in a higher revenue even when usingdelayed information, as apparent from Fig. 5. Next, Fig. 6a andFig. 6b shows the performance of the proposed asynchronousalgorithm for different network sizes and maximum delayvalues.

VI. CONCLUSION

We considered multi-agent stochastic optimization problemswhere the hypothesis that all agents are trying to learn commonparameters may be violated, with the additional stipulation thatagents do not even operate on a synchronized time-scale. Tosolve this problem such that agents give preference to locallyobserved information while incorporating the relevant infor-mation of others, we formulated this task as a decentralizedstochastic program with convex proximity constraints whichincentivize distinct nodes to make decisions which are close toone another. We derived an asynchronous stochastic variant ofthe Arrow-Hurwicz saddle point method to solve this problemthrough the use of a dual-augmented Lagrangian.

We established that in expectation, under a constant step-size regime, the time-average suboptimality and constraint vi-olation are contained in a neighborhood whose radius vanisheswith increasing number of iterations (Theorem 1). This resultextends existing results for multi-agent convex stochastic pro-grams with inequality constraints to asynchronous computingarchitectures which are important for wireless systems.

We then considered an empirical evaluation for the taskof pricing-based interference management in large distributedwireless networks. For this application setting, we observe thatthe theoretical convergence rates translate into practice, andfurther that the use of asynchronous updates breaks the compu-tational bottleneck associated with requiring devices to operateon a common clock. In particular, the asynchronous saddlepoint method learns slightly more slowly its synchronouscounterpart while yielding a substantial complexity reductionon a real wireless application, and thus holds promise for otheronline multi-agent settings where synchronous operations andconsensus are overly restrictive.

APPENDIX 0: DERIVATION OF GENERALIZED ALGORITHM

Consider the generalized the problem in (8). The stochasticaugmented Lagrangian takes the form

Lt(x,λ) =

N∑i=1

[f i(xi,θit)

+

(〈λi,hi

({xj ,θjt}j∈n′

i

)〉− δε

2

∥∥λi∥∥2)]. (35)

The primal update for the generalized asynchronous stochasticsaddle point algorithm at each node i is given by

xit+1 = PX[xit − ε

(∇xif i(xi[t]i ,θ

i[t]i)

+∑k∈n′

i

〈λkt ,∇xihk

({xj[t]j ,θ

j[t]j}j∈n′

k

)〉)]. (36)

through applying comparable logic to Section III. Likewise,the dual variable updates at each node i is given by

λit+1 =[(1− ε2δ)λit + ε

(hi({xj[t]j ,θ

j[t]j}j∈n′

i

)) ]+. (37)

Before proceeding with the convergence analysis, to have atractable derivation, let us define the compact notation for theprimal and dual delayed gradient as follows

∇xiLi[t]({xj[t]j ,λ

jt}j∈n′

i

):=(∇xif i(xi[t]i ,θ

i[t]i)

+∑k∈n′

i

〈λkt ,∇xihk

({xj[t]j ,θ

j[t]j}j∈n′

k

)〉)

(38)

which generalizes (13) to the setting of (8) and the notation

∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

):= hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)(39)

follows from (14). These generalized algorithm updates areused to simplify the notation in the subsequent analysis.

APPENDIX A: PROOF OF LEMMA 1

Consider the squared 2-norm of the difference between theiterate xit at time t+1 and an arbitrary feasible point xi ∈ XNand use (36) to express xit+1 in terms of xit,

‖xit+1−xi‖2 =‖PX [xit−ε∇xiLi[t]({xj[t]j ,λ

jt}j∈n′

i

)]−xi‖2.

(40)

where, we have utilized the compact notation defined in (38) tosubstitute (36) into (40). Since xi ∈ X , utilizing non-expansive

Page 11: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

11

Iterations, t0 200 400 600 800 1000

1 t

∑t u=1F(x

u)

101

102

103M=5M=10M=20

(a) Objective vs. iteration t, fixed τmax = 10.

Iterations, t200 400 600 800 1000

1 t

∑t u=1F(x

u)

30

35

40

45 τmax = 0τmax = 20τmax = 40

(b) Objective vs. iteration t, fixed M = 10.

Fig. 6: Performance analysis for different network size and delay values.

property of the projection operator in (40) and expanding thesquare

‖xit+1 − xi‖2≤‖xit − ε∇xiLi[t]({xj[t]j ,λ

jt}j∈n′

i

)− xi‖2

=‖xit − xi‖2

− 2ε〈∇xLi[t]({xj[t]j ,λ

jt}j∈n′

i

), (xit − xi)〉

+ ε2‖∇xLi[t]({xj[t]j ,λ

jt}j∈n′

i

)‖2. (41)

We reorder terms of the above expression such that thegradient inner product is on the left-hand side and then takesummation over i ∈ V , yielding Take the summation overnodes i ∈ V on both sides, we get

N∑i=1

〈∇xLi[t]({xj[t]j ,λ

jt}j∈n′

i

), (xit − xi)〉

≤ 1

N∑i=1

(‖xit − xi‖2−‖xit+1 − xi‖2

)+ε

2

N∑i=1

∥∥∥∇xiLi[t]({xj[t]j ,λ

jt}j∈n′

i

)∥∥∥2 . (42)

Let us use the following notation,

x[t] :=[x1[t]1

; · · ·xN[t]N], and λt :=

[λ1t ; · · · ;λNt

]. (43)

Utilizing this notation, we can write

〈∇xL[t](x[t],λt), (xt − x)〉

≤ 1

(‖xt−x‖2−‖xt+1−x‖2

)+ε

2‖∇xL[t](x[t],λt)‖2. (44)

Add and subtract 〈∇xL[t](x[t],λt),x[t]〉 to left hand side of(44) to obtain

〈∇xL[t](x[t],λt), (x[t] − x)〉

≤ 1

(‖xt − x‖2−‖xt+1 − x‖2

)+ε

2

∥∥∥∇xL[t](x[t],λt)∥∥∥2

+ 〈∇xL[t](x[t],λt), (x[t] − xt)〉. (45)

Observe now that since the functions f i(xi,θit) andhi({xj ,θjt}j∈n′

i) are convex with respect to optimization

variables for any given realization of the associated randomvariables, therefore the stochastic Lagrangian is a convexfunction of xi and xj [cf. (9)]. Hence, from the first order

convex inequality and the definition of Lagrangian in (10), itholds that

L[t](x[t],λt)−L[t](x,λt)≤〈∇xL[t](x[t],λt),(x[t]−x)〉. (46)

Substituting the upper bound in (45) into the right hand sideof (46) yields

L[t](x[t],λt)− L[t](x,λt)

≤ 1

(‖xt − x‖2−‖xt+1 − x‖2

)+ε

2

∥∥∥∇xL[t](x[t],λt)∥∥∥2

+ 〈∇xL[t](x[t],λt), (x[t] − xt)〉. (47)

We set this analysis aside and proceed to repeat the steps in(40)-(47) for the distance between the iterate λit+1 at time t+1and an arbitrary multiplier λi.

‖λit+1−λi‖2=‖[λit+ε∇λiLi[t]

({xj[t]j ,λ

jt}j∈n′

i

)]+−λi‖2 (48)

where we have substituted (37) to express λit+1 in terms of λit.Using the non-expansive property of the projection operator in(48) and expanding the square, we obtain

‖λit+1 − λi‖2 ≤ ‖λit + ε∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

)− λi‖2

= ‖λit − λ‖2

+ 2ε〈∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

), (λit − λi)〉

+ ε2‖∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

)‖2. (49)

Reorder terms in the above expression such that the gradientinner product term is on the left-hand side as

〈∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

), (λit − λi)〉

≥ 1

(‖λit+1 − λi‖2 − ‖λit − λi‖2

)− ε

2‖∇λiLi[t]

({xj[t]j ,λ

jt}j∈n′

i

)‖2. (50)

Page 12: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

12

Take the summation over nodes i ∈ V so that we may write

N∑i=1

〈∇λiLi[t]({xj[t]j ,λ

jt}j∈n′

i

), (λit − λi)〉

≥ 1

N∑i=1

(‖λit+1 − λi‖2 − ‖λit − λi‖2

)−

N∑i=1

ε

2‖∇λiLi[t]

({xj[t]j ,λ

jt}j∈n′

i

)‖2. (51)

Utilizing the notation defined in (43), we write (51) as follows

〈∇λL[t](x[t],λt), (λt − λ)〉

≥ 1

(‖λt+1−λ‖2−‖λt−λ‖2

)− ε

2‖∇λL[t](x[t],λt)‖2. (52)

Note that the online Lagrangian [cf. (9)] is a concave functionof its Lagrange multipliers, which implies that instantaneousLagrangian differences for fixed x[t] satisfy

L[t](x[t],λt)−L[t](x[t],λ) ≥ ∇λL[t](x[t],λt)T (λt−λ). (53)

By using the lower bound stated in (52) for the right handside of (53), we can write

L[t](x[t],λt)−L[t](x[t],λ) ≥ 1

(‖λt+1−λ‖2−‖λt − λ‖2

)− ε

2‖∇λL[t](x[t],λt)‖2. (54)

We now turn to establishing a telescopic property of theinstantaneous Lagrangian by combining the expressions in(47) and (54). To do so observe that the term L[t](x[t],λt)appears in both inequalities. Thus, subtracting the inequality(54) from those in (47) followed by reordering terms yieldsthe required result in (24). �

APPENDIX B: PROOF OF LEMMA 2

We first consider the expression in (24), and expand the left-hand side using the definition of the augmented Lagrangian in(10). Doing so yields the following expression,

N∑i=1

[f i(xi[t]i ,θi[t]i)− f

i(xi,θi[t]i)] +δε

2(‖λt‖2−‖λ‖2)

+

N∑i=1

[〈λi,hi

({xj[t]j ,λ

jt}j∈n′

i

)〉−〈λit,hi

({xj ,λjt}j∈n′

i

)〉]

≤ 1

(‖xt−x‖2−‖xt+1−x‖2 + ‖λt − λ‖2−‖λt+1−λ‖2

)+ε

2

(∥∥∥∇λL[t](x[t],λt)∥∥∥2 + ‖∇xL[t](x[t],λt)‖2

)+ 〈∇xL[t](x[t],λt), (x[t] − xt)〉. (55)

Let Ft denotes the sigma field collecting the algorithm historywhich collects the information for all random quantities as{θu,xu,λu}u<t. Note that this notation is slightly differentfrom the standard notation in literature and the maximum valueu can take is t − 1 which is for the synchronous case. Notethat the conditional expectation of the following term for givensigma algebra F[t] is equal to

E[ N∑i=1

[f i(xi[t]i ,θi[t]i)−f

i(xi,θi[t]i)] | F[t]

]=F (x[t])−F (x)

(56)

Taking the total expectation of (55) and utilizing the simplifiedexpression in (56), we get

E[F (x[t])− F (x)] +δε

2E(‖λt‖2−‖λ‖2)

+

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉−〈λit,hi

({xj ,θj[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖xt−x‖2−‖xt+1−x‖2 + ‖λt − λ‖2−‖λt+1−λ‖2

)+ε

2

(E∥∥∥∇λL[t](x[t],λt)

∥∥∥2 + E‖∇xL[t](x[t],λt)‖2)

+ E[〈∇xL[t](x[t],λt), (x[t] − xt)〉

]. (57)

Let us develop the upper bound on the termE[〈λit,hi

({xj ,θj[t]j}j∈n′

i

)〉]. Note that

E[〈λit,hi

({xj ,θj[t]j}j∈n′

i

)〉]

=E[E[〈λit,hi

({xj ,θj[t]j}j∈n′

i

)〉|F[t]

]]= E

[〈λit,E

[(hi({xj ,θj[t]j}j∈n′

i

))| F[t]

]〉]≤0, (58)

where the first equality in (58) holds from the law of iteratedaverages. The second equality holds since λit is deterministicfor given F[t], and the third is due to the fact that for anyfeasible {xj}j∈n′

i, E[hi

({xj ,θj[t]j}j∈n′

i

)] ≤ 0 due to the

Slater’s conditions (Assumption 2), where here we use thegeneralized constraint definition given in (4) which subsumes(5). Now, let’s use (58) in the left hand side of (57) to obtain

E[F (x[t])− F (x)] +δε

2E(‖λt‖2−‖λ‖2) (59)

+

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖xt−x‖2−‖xt+1−x‖2 + ‖λt − λ‖2−‖λt+1−λ‖2

)+ε

2

(E∥∥∥∇λL[t](x[t],λt)

∥∥∥2 + E‖∇xL[t](x[t],λt)‖2)

+ I,

where in (59) we have defined I on the right-hand side as

I := 〈∇xL[t](x[t],λt), (x[t] − xt)〉. (60)

Observe that (60) is the directional error of the stochasticgradient caused by asynchronous updates. To proceed further,an upper bound on the term I is required, which we derivenext. Using Cauchy Schwartz inequality, we obtain

I ≤∥∥∥∇xL[t](x[t],λt)

∥∥∥∥∥(x[t] − xt)∥∥ . (61)

Since the maximum delay is τ , we can write the differenceas∥∥xt − x[t]

∥∥ as sum of τ intermediate difference usingtriangular inequality as follows

∥∥xt−x[t]

∥∥≤ t−1∑s=t−τ

‖xs+1−xs‖≤εt−1∑s=t−τ

∥∥∇xL[s](x[s],λs)∥∥ . (62)

Page 13: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

13

where the inequality in (62) follows from the primal update in(36). For brevity, let us denote Bt :=

∥∥∇xL[t](x[t],λt)∥∥. Note

that in the definition of Bt, only t index is emphasized sincethe dual variable involved is λt. Substituting upper boundobtained in (62) into (61), simplyfying the notation using thedefinition of Bt and then taking expectation, we get

E [I] ≤ εt−1∑s=t−τ

E [BtBs] ≤ε

2

t−1∑s=t−τ

E[B2t +B2

s

]=ε

2

[τ · E

[B2t

]+

t−1∑s=t−τ

E[B2s

] ]. (63)

The second inequality in (63) follows directly using ab ≤a2+b2

2 . The last equality of (63) is obtained by expanding thesummation. Next, applying the gradient norm square upperbound of (22) and (23) into (63), we obtain

E [I] ≤ ε2

[ (2τ(N +M2)L2(1 + E

[‖λt‖2

]))

+

t−1∑s=t−τ

(2(N +M2)L2(1 + E[‖λs‖2])

) ]. (64)

Rearranging and collecting the like terms, we get

E [I] ≤ε[2τ(N +M2)L2 + τ(N +M2)L2E

[‖λt‖2

]+(N +M2)L2

t−1∑s=t−τ

E[‖λs‖2]]. (65)

Multiplying the last term of (65) by τ , we get

E [I] ≤ε[2τ(N +M2)L2 + τ(N +M2)L2E[‖λt‖2]

+ τ(N +M2)L2t−1∑s=t−τ

E[‖λs‖2]]

(66)

Let us define K1 := (N+M2)L2, expression in (66) can beexpressed as

E [I] ≤ ετK1

[2 +

t∑s=t−τ

E[‖λs‖2]]. (67)

Substitute the bounds developed in (22), (23) and (67) backinto (59), we get

E[F (x[t])− F (x)] +δε

2E(‖λt‖2−‖λ‖2)

+

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖xt−x‖2−‖xt+1−x‖2+‖λt−λ‖2−‖λt+1−λ‖2

)+ε

2

(2Mσ2

λ+2δ2ε2E[‖λt‖2

]+2(N+M2)L2(1 + E[‖λt‖2])

)+ ετK1

[2 +

t∑s=t−τ

E[‖λs‖2]]. (68)

Collecting the like terms together and defining K2 := Mσ2λ +

(N+M2)L2+τK1 and K3 :=δ2ε2+(N+M2)L2, we get

E[F (x[t])− F (x)] +δε

2E(‖λt‖2−‖λ‖2)

+

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖xt−x‖2−‖xt+1−x‖2 + ‖λt − λ‖2−‖λt+1−λ‖2

)+ ε[K2+K3E[‖λt‖2]+

τK1

2

t∑s=t−τ

E[‖λs‖2]]. (69)

Adding E[F (x[t])− F (xt)

]to the both sides of (69) and

apply Lipschitz continuity of the objective, yields

E[F (xt)− F (x)] +δε

2E(‖λt‖2−‖λ‖2)

+

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

(70)

≤ 1

2εE(‖xt−x‖2−‖xt+1−x‖2+‖λt−λ‖2−‖λt+1−λ‖2

)+ε[K2+K3E[‖λt‖2]+

τK1

2

t∑s=t−τ

E[‖λs‖2]]+LfE

[∥∥xt−x[t]

∥∥] .Now, we proceed to analyze the resulting termLfE

[∥∥xt−x[t]

∥∥]. Further note that using [E [X]]2 ≤ E[X2]

for any random variable X , we can write

E[∥∥xt−x[t]

∥∥]≤√E[∥∥xt−x[t]

∥∥2]≤(E[(t−1∑s=t−τ‖xk+1−xk‖

)2])1/2(71)

Inequality in (71) follows from the triangular inequality usingcomparable analysis to that which yields (62). Further utilizing

the result( U∑i=1

ai

)2≤ U

U∑i=1

a2i , we get

E[∥∥xt − x[t]

∥∥] ≤(τ

t−1∑s=t−τ

E[‖xk+1 − xk‖2

])1/2. (72)

Now using the upper bound for single iterate different in termsof gradient as in (62), we get

E[∥∥xt − x[t]

∥∥]≤(τε2

t−1∑s=t−τ

E[∥∥∇xL[s](x[s],λs)

∥∥2])1/2

≤ε(2τ(N+M2)L2

t−1∑s=t−τ

[1 + E[‖λs‖2]])1/2

.

(73)

Inequality in (73) holds due to application of gradient normbounds. From the standard inequality of

√1 + Z ≤ (1 + Z)

for all Z ≥ 0, we can write

E[∥∥xt − x[t]

∥∥] ≤ 2ετ√K1

t−1∑s=t−τ

[1 + E[‖λs‖2]]. (74)

Page 14: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

14

where we pull 2τ out of square root because the product iseither zero or grater than 1. Utilizing the upper bound of (74)for the last term in right hand side of (70) and taking thesummation over t = 1 to T , we get

T∑t=1

E[F (xt)− F (x)] +

T∑t=1

δε

2E(‖λt‖2−‖λ‖2)

+

T∑t=1

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖x1−x‖2+ ‖λ1 − λ‖2

)+ ε[TK2 +K3

T∑t=1

E[‖λt‖2] +τK1

2

T∑t=1

t∑s=t−τ

E[‖λs‖2]]

+ 2ετLf√K1

T∑t=1

t−1∑s=t−τ

[1 + E[‖λs‖2]] (75)

In (75), we exploit the telescopic property of the summandover differences in the magnitude of primal and dual iteratesto a fixed primal-dual pair (x,λ) which appears as the firstterm on right-hand side of (76), and the fact that the resultingexpression is deterministic. By assuming the dual variableis initialized as λ1 = 0 and then combining the like termstogether, we can upper bound the right hand side of (75),yielding

T∑t=1

E[F (xt)− F (x)] +

T∑t=1

δε

2E(‖λt‖2−‖λ‖2)

+

T∑t=1

N∑i=1

E[〈λi,hi

({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

≤ 1

2εE(‖x1−x‖2+ ‖λ‖2

)+ εTK2 + εK3

T∑t=1

E[‖λt‖2] + 2ετTLf√K1

+ ετ(K1

2+ 2Lf

√K1)

T∑t=1

t∑s=t−τ

[E[‖λs‖2]]. (76)

Note that in (76), in order to gather terms, an extra ‖λt‖2 isadded to the right-hand side. We upper bound the last term onthe right-hand side of (76) by considering

T∑t=1

t∑s=t−τ

‖λs‖2=‖λ1‖2 + (‖λ1‖2 + ‖λ2‖2)

+(‖λ1‖2 + ‖λ2‖2 + ‖λ3‖2) + · · ·+(‖λT−τ‖2+‖λT−τ+1‖2+· · ·+‖λT ‖2). (77)

The relationship in (77) then simplifies to

T∑t=1

t∑s=t−τ

‖λs‖2 ≤ (τ + 1)

T∑t=1

E[‖λt‖2

]. (78)

Utilizing this on the right hand side of (76), we get

E[ T∑t=1

[F (xt)−F (x)]+

N∑i=1

〈λi,T∑t=1

hi({xj[t]j ,θ

j[t]j}j∈n′

i

)〉

−(δεT

2+

1

)‖λ‖2

]≤ 1

2ε‖x1−x‖2 +

εTK

2

+ (ε/2)(K4−δ)T∑t=1

E[‖λt‖2]. (79)

where K := 2K2 + 4τLf√K1 and K4 :=

(2K3+(τ + 1)τ(K1+ 4Lf√K1)). Now selecting δ such that

(K4 − δ) ≤ 0 makes the last term on the right-hand side ofthe preceding expression null, so that we may write

E[ T∑t=1

[F (xt)−F (x)]+

N∑i=1

[〈λi,

T∑t=1

hi({xj[t]j ,θ

j[t]j}j∈n′

i

)〉]

−(δεT

2+

1

)‖λ‖2

]≤ 1

2ε‖x1−x‖2 +

εTK

2. (80)

Observe that Lemma 2 is a special case of (80) with simplifiedconstraint functions that only allow for pair-wise coupling ofthe decisions of distinct nodes (see Remark 1). �

APPENDIX C: PROOF OF THEOREM 1

At this point, we note that the left-hand side of the expres-sion in (80), and hence (25), consists of three terms. The firstis the accumulation over time of the global sub-optimality,which is a sum of all local losses at each node as defined in(2); the second is the inner product of the an arbitrary Lagrangemultiplier λ with the time-aggregation of constraint violation;and the last depends on the magnitude of this multiplier. Wemay use these later terms to define an “optimal” Lagrangemultiplier to control the growth of the long-term constraintviolation of the algorithm. This technique is inspired by theapproach in [43], [52]. To do so, define the augmented dualfunction g(λ) using the later two terms on the left-hand sideof (79)

g(λ)=

N∑i=1

〈λi,T∑t=1

hi({xj[t]j ,θ

j[t]j}j∈n′

i

)〉−(δεT

2+

1

)‖λ‖2.

Computing the gradient and solving the resulting stationaryequation over the range RM+ yields

λi

= Z(ε)[ T∑t=1

hi({xj[t]j ,θ

j[t]j}j∈n′

i

)]+

(81)

for all (i, z) ∈ E , where Z(ε) := 1(Tδε+1/ε) . Substituting the

selection λi = λi

defined by (81) into (80) results in thefollowing expression

E

[T∑t=1

[F (xt)−F (x)]+Z(ε)

N∑i=1

∥∥∥∥∥[T∑t=1

(hi({xj[t]j,θ

j[t]j}j∈n′

i

))]+

∥∥∥∥∥2]

≤ 1

2ε‖x1−x‖2 +

εTK

2≤√T

2

(‖x1−x‖2+K

). (82)

The second inequality in (82) is obtained by selecting theconstant step-size ε=1/

√T . This result allows us to derive

Page 15: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

15

both the convergence of the global objective and the feasibilityof the stochastic saddle point iterates.

We first consider the average objective error sequenceE[F (xt) − F (x∗)]. To do so, subtract the last term on theleft-hand side of (82) from both sides, and note that theresulting term is non-positive. This observation allows us toomit the constraint slack term in (82), which taken with theselection x = x∗ [cf. (3)] and pulling the expectation insidethe summand, yields

T∑t=1

E[(F (xt)−F (x∗))]≤√T

2

(‖x1−x∗‖2+K

)= O(

√T ),

which is as stated in (27). Now we turn to establishing asublinear growth of the constraint violation in T , using theexpression in (82). Note that from the Lipschitz continu-ity of the objective function, we have |F (xt)−F (x?)| ≤Lf ‖xt−x?‖. An immediate consequence of this inequalityis that F (xt)−F (x?)≥−2LfR, using this in (82) yields

E

[1√

T (δ + 1)

N∑i=1

∥∥∥∥∥[T∑t=1

(hi({xj[t]j ,θ

j[t]j}j∈n′

i

))]+

∥∥∥∥∥2]

≤√T

2

(‖x1−x∗‖2+K

)+ 2TLfR. (83)

which, after multiplying both sides by 2√T (δ + 1) yields

E

[N∑i=1

∥∥∥∥∥[T∑t=1

(hi({xj[t]j ,θ

j[t]j}j∈n′

i

))]+

∥∥∥∥∥2]

(84)

≤(

2√T (δ + 1)

)(√T2

(‖x1−x∗‖2+K

)+ 2TLfR

).

We complete the proof by noting that the squareof the network-in-aggregate constraint violation∑Ni=1

∥∥∥∥[∑Tt=1

(hi({xj[t]j ,θ

j[t]j}j∈n′

i

))]+

∥∥∥∥2 upper bounds

the square of individual proximity constraint violations sinceit is a sum of positive squared terms, i.e.,

E

[N∑i=1

∥∥∥∥∥[T∑t=1

(hi({xj[t]j ,θ

j[t]j}j∈n′

i

))]+

∥∥∥∥∥2]

≥E

[[ T∑t=1

(hij(x

i[t]i,xj[t]j ,θ

i[t]i ,θ

j[t]j

)−γij) ]2

+

](85)

where we utilized the notation defined in (5) and the inequalitythat the norm square of a vector is always greater than squareof each element of the vector. The inequality in (85) is truefor any arbitrary ij in the right hand side.

Thus the right-hand side of (85) may be used in place ofthe left-hand side of (84), implying that

E[[ T∑t=1

hij

(xi[t]i ,x

j[t]j,θi[t]i ,θ

j[t]j

) ]2+

](86)

≤(

2√T (δ + 1)

)(√T2

(‖x1−x∗‖2+K

)+ 2TLfR

).

In order to present the results for the special case discussedin (3), compute the square root of both sides of (86) and takethe summation over all (i, j) ∈ E to conclude (28). �

REFERENCES

[1] A. S. Bedi, A. Koppel, and K. Rajawat, “Asynchronous saddle pointmethod: Interference management through pricing,” IEEE CDC 2018(submitted).

[2] L. Chen, S. Low, M. Chiang, and J. Doyle, “Cross-Layer CongestionControl, Routing and Scheduling Design in Ad Hoc Wireless Networks,”in in Proc. IEEE INFOCOM, April 2006, pp. 1–13.

[3] A. Koppel, J. Fink, G. Warnell, E. Stump, and A. Ribeiro, “Onlinelearning for characterizing unknown environments in ground roboticvehicle models,” in Proc. of IEEE IROS, 2016, pp. 626–633.

[4] D. Palomar and M. Chiang, “A tutorial on decomposition methods fornetwork utility maximization,” IEEE J. Sel. Areas Commun., vol. 24,no. 8, pp. 1439–1451, Aug 2006.

[5] S. Shalev-Shwartz, “Online learning and online convex optimization,”Found. Trends Mach. Learn., vol. 4, no. 2, pp. 107–194, Feb. 2012.

[6] J. Tsitsiklis, D. Bertsekas, and M. Athans, “Distributed asynchronousdeterministic and stochastic gradient optimization algorithms,” IEEETrans. Autom. Control, vol. 31, no. 9, pp. 803–812, 1986.

[7] V. N. Vapnik, The Nature of Statistical Learning Theory. New York:Springer-Verlag, 1995.

[8] O. Bousquet and L. Bottou, “The tradeoffs of large scale learning,” inAdv. Neural Inf. Process. Syst., 2008, pp. 161–168.

[9] H. Robbins and S. Monro, “A stochastic approximation method,” Ann.Math. Statist., vol. 22, no. 3, pp. 400–407, 09 1951.

[10] L. Bottou, “Online algorithms and stochastic approximations,” in OnlineLearning and Neural Networks, D. Saad, Ed. Cambridge, UK:Cambridge University Press, 1998.

[11] A. Nedic and A. Ozdaglar, “Distributed subgradient methods for multi-agent optimization,” IEEE Trans. Autom. Control, vol. 54, no. 1, 2009.

[12] K. I. Tsianos and M. G. Rabbat, “Distributed dual averaging for convexoptimization under communication delays,” in Proc. of IEEE ACC, 2012,2012, pp. 1067–1072.

[13] A. Nedic and A. Ozdaglar, “Subgradient methods for saddle-pointproblems,” J. Optim. Theory Appl., vol. 142, no. 1, pp. 205–228, 2009.

[14] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributedoptimization and statistical learning via the alternating direction methodof multipliers,” Found. Trends Mach. Learn., vol. 3, no. 1, 2011.

[15] A. Koppel, F. Jakubiec, and A. Ribeiro, “A saddle point algorithm fornetworked online convex optimization,” IEEE Trans. Signal Process.,p. 15, Oct 2015.

[16] M. G. Rabbat and K. I. Tsianos, “Asynchronous decentralized optimiza-tion in heterogeneous systems,” in Decision and Control (CDC), 2014IEEE 53rd Annual Conference on. IEEE, 2014, pp. 1125–1130.

[17] D. P. Bertsekas and J. N. Tsitsiklis, Parallel and distributed computation:numerical methods. Prentice hall Englewood Cliffs, NJ, 1989, vol. 23.

[18] B. Recht, C. Re, S. Wright, and F. Niu, “Hogwild: A lock-free approachto parallelizing stochastic gradient descent,” in Adv. Neural Inf. Process.Syst., 2011, pp. 693–701.

[19] A. Koppel, B. M. Sadler, and A. Ribeiro, “Proximity without consensusin online multiagent optimization,” IEEE Trans. Signal Process., vol. 65,no. 12, pp. 3062–3077, June 2017.

[20] A. Koppel, G. Warnell, E. Stump, and A. Ribeiro, “D4l: Decentralizeddynamic discriminative dictionary learning,” IEEE Trans. Signal Inf.Process. over Netw., 2017.

[21] A. S. Bedi and K. Rajawat, “Asynchronous incremental stochastic dualdescent algorithm for network resource allocation,” IEEE Trans. SignalProcess., vol. 66, no. 9, pp. 2229–2244, May 2018.

[22] K. Arrow, L. Hurwicz, and H. Uzawa, Studies in linear and non-linearprogramming, ser. Stanford Mathematical Studies in the Social Sciences.Stanford University Press, Stanford, Dec. 1958, vol. II.

[23] R. Caruana, “Multitask learning,” Mach. Learn., vol. 28, no. 1, pp. 41–75, Jul. 1997.

[24] J. Chen, C. Richard, and A. H. Sayed, “Multitask diffusion adaptationover networks,” IEEE Trans. Signal Process., vol. 62, no. 16, pp. 4129–4144, 2014.

[25] K. Srivastava and A. Nedic, “Distributed asynchronous constrainedstochastic optimization,” Selected Topics in Signal Processing, IEEEJournal of, vol. 5, no. 4, pp. 772–790, 2011.

[26] A. H. Sayed and X. Zhao, “Asynchronous adaptive networks,” arXivpreprint arXiv:1511.09180, 2015.

Page 16: koppel.netlify.comkoppel.netlify.com/assets/papers/2017_bedi_etal.pdf1 Asynchronous Saddle Point Algorithm for Stochastic Optimization in Heterogeneous Networks Amrit Singh Bedi?,

16

[27] J. N. Tsitsiklis, “Asynchronous stochastic approximation and Q-learning,” Machine learning, vol. 16, no. 3, pp. 185–202, 1994.

[28] M. J. Neely, “Asynchronous control for coupled markov decision sys-tems,” in Proc. of the IEEE Inf. Theory Workshop, 2012, pp. 287–291.

[29] A. Agarwal and J. C. Duchi, “Distributed delayed stochastic optimiza-tion,” in Adv. Neural Inf. Process. Syst., 2011, pp. 873–881.

[30] S. Zheng, Q. Meng, T. Wang, W. Chen, N. Yu, Z.-M. Ma, and T.-Y. Liu,“Asynchronous stochastic gradient descent with delay compensation,” inInternational Conference on Machine Learning, 2017, pp. 4120–4129.

[31] T.-H. Chang, M. Hong, W.-C. Liao, and X. Wang, “Asynchronousdistributed ADMM for large-scale optimization-part i: Algorithm andconvergence analysis,” arXiv preprint arXiv:1509.02597, 2015.

[32] R. Zhang and J. Kwok, “Asynchronous distributed ADMM for consensusoptimization,” in Proc. of the ICML), 2014, pp. 1701–1709.

[33] E. Wei and A. Ozdaglar, “On the o (1/k) convergence of asynchronousdistributed alternating direction method of multipliers,” arXiv preprintarXiv:1307.8254, 2013.

[34] Q. Ling, W. Shi, G. Wu, and A. Ribeiro, “Dlm: Decentralized linearizedalternating direction method of multipliers,” IEEE Transactions onSignal Processing, vol. 15, no. 63, pp. 4051–4064, 2015.

[35] A. Mokhtari, W. Shi, Q. Ling, and A. Ribeiro, “Decentralized quadrat-ically approximated alternating direction method of multipliers,” inSignal and Information Processing (GlobalSIP), 2015 IEEE GlobalConference on. IEEE, 2015, pp. 795–799.

[36] D. Jakovetic, J. M. F. Xavier, and J. M. F. Moura, “Fast distributedgradient methods,” CoRR, vol. abs/1112.2972, Apr. 2011.

[37] S. Ram, A. Nedic, and V. Veeravalli, “Distributed stochastic subgradientprojection algorithms for convex optimization,” J Optimiz. Theory App.,vol. 147, no. 3, pp. 516–545, Sep. 2010.

[38] K. Yuan, Q. Ling, and W. Yin, “On the convergence of decentralizedgradient descent,” ArXiv e-prints 1310.7063, Oct. 2013.

[39] Z. J. Towfic and A. H. Sayed, “Stability and performance limits ofadaptive primal-dual networks,” IEEE Trans. Signal Process., vol. 63,no. 11, pp. 2888–2903.

[40] M. Rabbat, R. Nowak, and J. Bucklew, “Generalized consensus compu-tation in networked systems with erasure links,” in 6th IEEE SPAWC,Jun. 5-8 2005, pp. 1088–1092.

[41] F. Jakubiec and A. Ribeiro, “D-map: Distributed maximum a posterioriprobability estimation of dynamic systems,” IEEE Trans. Signal Pro-cess., vol. 61, no. 2, pp. 450–466, Feb. 2013.

[42] E. Wei and A. Ozdaglar, “Distributed alternating direction method ofmultipliers,” in 51st IEEE CDC. IEEE, 2012, pp. 5445–5450.

[43] M. Mahdavi, R. Jin, and T. Yang, “Trading regret for efficiency: onlineconvex optimization with long term constraints,” J. Mach. Learn. Res.,vol. 13, no. Sep, pp. 2503–2528, 2012.

[44] D. P. Bertsekas, A. Nedic, and A. E. Ozdaglar, Convex Analysis andOptimization. Athena Scientific, 2003.

[45] A. Nedic, D. P. Bertsekas, and V. S. Borkar, “Distributed asynchronousincremental subgradient methods,” Studies in Computational Mathemat-ics, vol. 8, no. C, pp. 381–407, 2001.

[46] A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro, “Robust stochasticapproximation approach to stochastic programming,” SIAM J. Optimiz.,vol. 19, no. 4, pp. 1574–1609, 2009.

[47] F. R. Bach, “Adaptivity of averaged stochastic gradient descent to localstrong convexity for logistic regression.” J. Mach. Learn. Res., vol. 15,no. 1, pp. 595–627, 2014.

[48] M. Zinkevich, “Online convex programming and generalized infinitesi-mal gradient ascent,” in Proc. of the ICML, vol. 20, no. 2, WashingtonDC, USA, Aug. 21-24 2003, pp. 928–936.

[49] S. Haykin, “Cognitive radio: brain-empowered wireless communica-tions,” IEEE J. Sel. Areas Commun., vol. 23, no. 2, pp. 201–220, 2005.

[50] D. Lopez-Perez, A. Valcarce, G. De La Roche, and J. Zhang, “OFDMAfemtocells: A roadmap on interference avoidance,” IEEE Commun.Mag., vol. 47, no. 9, 2009.

[51] S. Bu, F. R. Yu, and H. Yanikomeroglu, “Interference-aware energy-efficient resource allocation for OFDMA-based heterogeneous networkswith incomplete channel state information,” IEEE Trans. Veh. Technol.,vol. 64, no. 3, pp. 1036–1050, 2015.

[52] R. Jenatton, J. Huang, and C. Archambeau, “Adaptive algorithms foronline convex optimization with long-term constraints,” in Proc. of theICML, 2016, pp. 402–411.

Amrit Singh Bedi (S’16) received the Diploma de-gree in electronics and communication engineering(ECE) from Guru Nanak Dev Polytechnic, Ludhiana,India, in 2009, the B.Tech. degree in ECE fromthe Rayat and Bahra Institute of Engineering andBio-Technology, Kharar, India, in 2012, the M.Tech.and a Ph.D. degree in electrical engineering (EE)from IIT Kanpur, in 2017 and 2018, respectively.Currently, he is a postdoctoral fellow with the U.S.Army Research laboratory, Adelphi, MD, USA. Heis currently involved in developing stochastic opti-

mization algorithms for supervised learning. His research interests includedistributed stochastic optimization for networks, time-varying optimization,and machine learning. His paper selected as a Best Paper Finalist at the 2017IEEE Asilomar Conference on Signals, Systems, and Computers.

Alec Koppel began as a Research Scientist atthe U.S. Army Research Laboratory in the Com-putational and Information Sciences Directorate inSeptember of 2017. He completed his Master’s de-gree in Statistics and Doctorate in Electrical andSystems Engineering, both at the University of Penn-sylvania (Penn) in August of 2017. He is also a par-ticipant in the Science, Mathematics, and Researchfor Transformation (SMART) Scholarship Programsponsored by the American Society of EngineeringEducation. Before coming to Penn, he completed his

Master’s degree in Systems Science and Mathematics and Bachelor’s Degreein Mathematics, both at Washington University in St. Louis (WashU), Mis-souri. His research interests are in the areas of signal processing, optimization,and learning theory. His current work focuses on optimization and learningmethods for streaming data applications, with an emphasis on problems arisingin autonomous systems. He co-authored a paper selected as a Best PaperFinalist at the 2017 IEEE Asilomar Conference on Signals, Systems, andComputers.

Ketan Rajawat (S’06-M’12) received his B.Techand M.Tech degrees in Electrical Engineering fromthe Indian Institute of Technology (IIT) Kanpur,India, in 2007, and his Ph.D. degree in Electricaland Computer Engineering from the University ofMinnesota, USA, in 2012. Since 2012, he has beenan Assistant Professor at the Department of Elec-trical Engineering, IIT Kanpur. His current researchfocuses on distributed algorithms, online learning,and control of networked systems. He is currentlyan Associate Editor of the IEEE Communications

Letters. He is a recipient of the Indian National Science Academy (INSA)Medal for Young Scientists in 2018 and the P. K. Kelkar Young FacultyResearch Fellowship in 2018.