248 ieee transactions on network science and …

13
Stochastic Subgradient Algorithms for Strongly Convex Optimization Over Distributed Networks Muhammed O. Sayin , N. Denizcan Vanli , Suleyman S. Kozat, Senior Member, IEEE, and Tamer Ba¸ sar, Life Fellow, IEEE Abstract—We study diffusion and consensus based optimization of a sum of unknown convex objective functions over distributed networks. The only access to these functions is through stochastic gradient oracles, each of which is only available at a different node; and a limited number of gradient oracle calls is allowed at each node. In this framework, we introduce a convex optimization algorithm based on stochastic subgradient descent (SSD) updates. We use a carefully designed time-dependent weighted averaging of the SSD iterates, which yields a convergence rate of O N ffiffiffi N p ð1sÞT after T gradient updates for each node on a network of N nodes, where 0 s < 1 denotes the second largest singular value of the communication matrix. This rate of convergence matches the performance lower bound up to constant terms. Similar to the SSD algorithm, the computational complexity of the proposed algorithm also scales linearly with the dimensionality of the data. Furthermore, the communication load of the proposed method is the same as the communication load of the SSD algorithm. Thus, the proposed algorithm is highly efficient in terms of complexity and communication load. We illustrate the merits of the algorithm with respect to the state-of-art methods over benchmark real life data sets. Index Terms—Distributed processing, convex optimization, online learning, diffusion strategies, consensus strategies Ç 1 INTRODUCTION T HE demand for large-scale networks consisting of multi- ple agents (i.e., nodes) [1] with different objectives is steadily growing due to their increased efficiency and scal- ability compared to centralized distributed structures [2], [3], [4], [5], [6]. A wide range of problems in the context of distributed and parallel processing can be considered as a minimization of a sum of objective functions, where each function (or information on each function) is available only to a single agent or node [7], [8], [9]. In such practical appli- cations, it is essential to process the information in a decen- tralized manner since transferring the objective functions as well as the entire resources (e.g., data) may not be feasible or possible [10], [11], [12], [13]. For example, in a distributed data mining scenario, privacy considerations may prohibit sharing of the objective functions [7], [8], [9]. Similarly, in a distributed wireless network, energy considerations may limit the communication rate between agents [14], [15], [16], [17]. In such settings, parallel or distributed processing algorithms, where each node performs its own processing and shares information subsequently, are preferable over the centralized methods [18], [19], [20], [21]. Here, we consider minimization of a sum of unknown convex objective functions, where each agent (or node) observes only its particular objective function via the sto- chastic gradient oracles. Particularly, we seek to minimize this sum of functions with a limited number of gradient ora- cle calls at each agent. In this framework, we introduce a dis- tributed online convex optimization algorithm based on stochastic subgradient descent (SSD) iterates that efficiently minimizes this cost function. Specifically, each agent uses a time-dependent weighted combination of the SSD iterates and achieves the presented performance guarantees, which matches the lower bounds presented in [22], only with a rela- tively small excess term caused by the unknown network model. The proposed method is comprehensive, in that any communication strategy, such as the diffusion [3] and the consensus [6] strategies, are incorporated into our algorithm in a straightforward manner as shown in the paper. We com- pare the performance of our algorithm with respect to the state-of-the-art methods [6], [11], [23] in the literature and present substantial performance improvements for various well-known network topologies and benchmark data sets. The distributed network framework is successfully used in wireless sensor networks [24], [25], [26], [27], [28], [29], as well as for convex optimization via projected subgradient techniques [6], [7], [8], [9], [10], [11]. In [11], the authors demonstrate the performance of the least mean squares (LMS) algorithm over distributed networks using different diffusion strategies. We emphasize that this problem can also be cast as a distributed convex optimization problem, and hence our results here can be applied to these problems M. O. Sayin and T. Ba¸ sar are with the Coordinated Science Laboratory, University of Illinois at Urbana-Champaign (UIUC), Urbana, IL 61801. E-mail: {sayin2, basar1}@illinois.edu. N. D. Vanli is with the Laboratory for Information and Decision Systems, Massachusetts Institute of Technology (MIT), Cambridge, MA 02139. E-mail: [email protected]. S. S. Kozat is with the Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey. E-mail: [email protected]. Manuscript received 12 Jan. 2016; revised 11 May 2017; accepted 1 June 2017. Date of publication 7 June 2017; date of current version 11 Dec. 2017. (Corresponding author: Muhammed O. Sayin.) Recommended for acceptance by J. Corte´s. For information on obtaining reprints of this article, please send e-mail to: [email protected], and reference the Digital Object Identifier below. Digital Object Identifier no. 10.1109/TNSE.2017.2713396 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017 2327-4697 ß 2017 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See ht_tp://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Upload: others

Post on 29-Apr-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

Stochastic Subgradient Algorithms for StronglyConvex Optimization Over Distributed Networks

Muhammed O. Sayin , N. Denizcan Vanli , Suleyman S. Kozat, Senior Member, IEEE,

and Tamer Basar, Life Fellow, IEEE

Abstract—We study diffusion and consensus based optimization of a sum of unknown convex objective functions over distributed

networks. The only access to these functions is through stochastic gradient oracles, each of which is only available at a different node;

and a limited number of gradient oracle calls is allowed at each node. In this framework, we introduce a convex optimization algorithm

based on stochastic subgradient descent (SSD) updates. We use a carefully designed time-dependent weighted averaging of the SSD

iterates, which yields a convergence rate of O NffiffiffiN

pð1�sÞT� �

after T gradient updates for each node on a network of N nodes, where

0 � s < 1 denotes the second largest singular value of the communication matrix. This rate of convergence matches the performance

lower bound up to constant terms. Similar to the SSD algorithm, the computational complexity of the proposed algorithm also scales

linearly with the dimensionality of the data. Furthermore, the communication load of the proposed method is the same as the

communication load of the SSD algorithm. Thus, the proposed algorithm is highly efficient in terms of complexity and communication

load. We illustrate the merits of the algorithm with respect to the state-of-art methods over benchmark real life data sets.

Index Terms—Distributed processing, convex optimization, online learning, diffusion strategies, consensus strategies

Ç

1 INTRODUCTION

THE demand for large-scale networks consisting of multi-ple agents (i.e., nodes) [1] with different objectives is

steadily growing due to their increased efficiency and scal-ability compared to centralized distributed structures [2],[3], [4], [5], [6]. A wide range of problems in the context ofdistributed and parallel processing can be considered as aminimization of a sum of objective functions, where eachfunction (or information on each function) is available onlyto a single agent or node [7], [8], [9]. In such practical appli-cations, it is essential to process the information in a decen-tralized manner since transferring the objective functions aswell as the entire resources (e.g., data) may not be feasibleor possible [10], [11], [12], [13]. For example, in a distributeddata mining scenario, privacy considerations may prohibitsharing of the objective functions [7], [8], [9]. Similarly, in adistributed wireless network, energy considerations maylimit the communication rate between agents [14], [15], [16],[17]. In such settings, parallel or distributed processing

algorithms, where each node performs its own processingand shares information subsequently, are preferable overthe centralized methods [18], [19], [20], [21].

Here, we consider minimization of a sum of unknownconvex objective functions, where each agent (or node)observes only its particular objective function via the sto-chastic gradient oracles. Particularly, we seek to minimizethis sum of functions with a limited number of gradient ora-cle calls at each agent. In this framework, we introduce a dis-tributed online convex optimization algorithm based onstochastic subgradient descent (SSD) iterates that efficientlyminimizes this cost function. Specifically, each agent uses atime-dependent weighted combination of the SSD iteratesand achieves the presented performance guarantees, whichmatches the lower bounds presented in [22], only with a rela-tively small excess term caused by the unknown networkmodel. The proposed method is comprehensive, in that anycommunication strategy, such as the diffusion [3] and theconsensus [6] strategies, are incorporated into our algorithmin a straightforwardmanner as shown in the paper.We com-pare the performance of our algorithm with respect to thestate-of-the-art methods [6], [11], [23] in the literature andpresent substantial performance improvements for variouswell-known network topologies and benchmark data sets.

The distributed network framework is successfully usedin wireless sensor networks [24], [25], [26], [27], [28], [29], aswell as for convex optimization via projected subgradienttechniques [6], [7], [8], [9], [10], [11]. In [11], the authorsdemonstrate the performance of the least mean squares(LMS) algorithm over distributed networks using differentdiffusion strategies. We emphasize that this problem canalso be cast as a distributed convex optimization problem,and hence our results here can be applied to these problems

� M. O. Sayin and T. Basar are with the Coordinated Science Laboratory,University of Illinois at Urbana-Champaign (UIUC), Urbana, IL 61801.E-mail: {sayin2, basar1}@illinois.edu.

� N. D. Vanli is with the Laboratory for Information and Decision Systems,Massachusetts Institute of Technology (MIT), Cambridge, MA 02139.E-mail: [email protected].

� S. S. Kozat is with the Department of Electrical and ElectronicsEngineering, Bilkent University, Ankara 06800, Turkey.E-mail: [email protected].

Manuscript received 12 Jan. 2016; revised 11 May 2017; accepted 1 June 2017.Date of publication 7 June 2017; date of current version 11 Dec. 2017.(Corresponding author: Muhammed O. Sayin.)Recommended for acceptance by J. Cortes.For information on obtaining reprints of this article, please send e-mail to:[email protected], and reference the Digital Object Identifier below.Digital Object Identifier no. 10.1109/TNSE.2017.2713396

248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

2327-4697� 2017 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See ht _tp://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

in a straightforward manner. In [10], the authors considerthe cooperative optimization of the cost function under con-vex inequality constraints. However, the problem formula-tion as well as the convergence results in this paper aresubstantially different from the ones in [10]. In particular, in[10], agents seek to minimize an approximation of the origi-nal optimization problem through penalty functions whilewe directly consider the original optimization problem. Fur-thermore, Reference [10] provides an upper bound on themean square error as the number of iterates goes to infinityfor sufficiently small step sizes. Yet that upper bound goesto zero as the step size goes to zero. On the other hand,here, we not only show that through the proposed approacheach agent achieves the minimum cost for a certain stepsize, but also provide an upper bound on the convergencerate, while that upper bound matches the lower bound pro-vided in [22] up to constant terms.

In [2], [6], the authors present a (constrained in [6] andunconstrained in [2]) deterministic analysis of the SSD iter-ates and our results build on them by illustrating a strongerconvergence bound in expectation while also providingMSD analyses of the SSD iterates. Similarly, a regret analysisis conducted for every possible input stream in an online anddistributed manner in [30] for general convex cost functions;and in [31] under Lipschitz continuous and strongly convexcost functions, where the latter achieves a regret diminishingat a faster rate of Oðlog ðT Þ=T Þ (after T iterates). On the con-trary, we study the distributed online convex optimizationproblem in the expectation sense (with respect to the datastatistics), i.e., not in an individual sequence manner, wherewe show that SSD iterates achieve the optimal convergencerate ofOð1=T Þ. In [7], [8], [9], the authors consider the distrib-uted convex optimization problem and present probability-1and mean square convergence results of the SSD iterates. Inthis paper, on the other hand, we provide the expected con-vergence rate of our algorithm and the mean square devia-tion (MSD) of the SSD iterates at any time instant.

Similar convergence analyses have recently been carriedout in the computational learning theory literature [22],[23], [32], [33], [34]. In [32], the authors provide determin-istic bounds on the learning performance (i.e., regret) of theSSD algorithm. In [33], these analyses are extended and aregret-optimal learning algorithm is proposed. Along simi-lar lines, in [23], the authors describe a method to make theSSD algorithm optimal for strongly convex optimization.However, these approaches rely on the smoothness of theoptimization problem. In [34], a different method to achievethe optimal convergence rate is proposed and its perfor-mance is analyzed. In this paper, however, convex optimi-zation is performed over a network of localized learners,unlike in [23], [32], [33], [34]. Our results entail convergencerates over any unknown communication graph, and in thissense build upon the analyses of the centralized learners.Furthermore, unlike [23], [33], our algorithm does notrequire the optimization problem to be sufficiently smooth.

Distributed convex optimization appears in a wide rangeof practical applications in wireless sensor networks and real-time control systems [3], [4], [5]. We introduce a comprehen-sive approach to this setup by proposing an online algorithm,whose expected performance is asymptotically the same asthe performance of the optimal centralized processor. Our

results are generic for any probability distribution on thedata, not necessarily Gaussian, unlike the conventional worksin the literature [11], [12]. Our experiments over different net-work topologies, various data sets and cost functions demon-strate the superiority and robustness of our approach withrespect to the state-of-the-art methods in the literature.

Ourmain contributions can thus be summarized as follows.

1) We introduce a distributed online convex optimiza-tion algorithm based on SSD iterates, which achieves

an optimal convergence rate of O NffiffiffiN

pð1�sÞT� �

after T gra-

dient updates, for each and every node of the net-work, where N is the number of nodes. Weemphasize that this convergence rate is optimal sinceit achieves the lower bounds presented in [22] up toconstant terms.

2) We show that MSD between the time weighted aver-age and the optimal solution is also upper bounded

by O NffiffiffiN

pð1�sÞT� �

after T gradient updates while MSD

between the average of the iterates (which can beattained if the agents continue to exchange informa-tion without gradient updates) and the optimal solu-

tion is upper bounded by OffiffiffiN

pð1�sÞT� �

.

3) Our analyses can be extended to analyze the perform-ances of the diffusion and consensus strategies in astraightforwardmanner as illustrated in the paper.

4) We demonstrate that the algorithm introduced out-performs the state-of-the-art methods in terms ofnormalized accumulated error and MSD from theoptimal solution under various network topologiesand benchmark data sets.

The organization of the paper is as follows. In Section 2,we introduce the distributed convex optimization frame-work and provide the notations. We then introduce themain result of the paper, i.e., an SSD based convex optimiza-tion algorithm, in Section 3 and analyze the convergencerate of the algorithm. In Section 4, we demonstrate the per-formance of our algorithm with respect to the state-of-the-art methods through simulations and then conclude thepaper with several remarks in Section 5.

2 PROBLEM FORMULATION

2.1 Notation and Preliminaries

Throughout the paper, all vectors are column vectors andrepresented by boldface lowercase letters. Matrices are rep-resented by boldface uppercase letters. For a matrix HH,HHj jj jF is the Frobenius norm. For a vector xx, xxj jj j ¼

ffiffiffiffiffiffiffiffiffixxTxx

pis

the ‘2-norm. 0 (and 1) denotes a vector with all zero (andone) elements and the dimensions can be understood fromthe context. For a matrix HH, HHij represents its entry at theith row and jth column.

For a non-empty, closed and convex set W � Rm, PWdenotes the Euclidean projection ontoW, i.e.,

PWðww0Þ ¼ argminww2W

ww� ww0j jj j: (1)

We say that a convex function (possibly non-smooth)f : Rm ! R on the convex domain W has the subgradientset @fð�Þ � Rm at a point ww0 2 W if

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 249

Page 3: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

gg 2 @fðww0Þ , fðwwÞ � fðww0Þ þ ggT ðww� ww0Þ 8ww 2 W:

Furthermore, we say that f is �-strongly convex on W if,and only if, for all ww;ww0 2 W and gg 2 @fðww0Þ, we have

fðwwÞ � fðww0Þ þ ggT ðww� ww0Þ þ �

2ww� ww0j jj j2: (2)

2.2 System Overview

Consider a static and connected network of N-agents withprocessing and communication capabilities. Over the net-work, each agent has connections with certain other agents,i.e., the ones in his/her neighborhood, and can exchangeinformation with them. We can represent such a networkthrough an undirected graph, where the vertices and theedges correspond to the agents and the communicationlinks between them, respectively, as seen in Fig. 1.

Each agent seeks to minimize F : Rm ! R, which is a sumof �-strongly convex (possibly non-smooth) local cost func-tions Fi : R

m ! R, for i ¼ 1; . . . ; N , i.e., each agent aims to

minww2W

F ðwwÞ ¼ minww2W

XNi¼1

FiðwwÞ; (3)

whereW � Rm is a non-empty, closed and convex set. How-ever, the cost function F is unknown to the agents, and eachagent i has only access to F via at most T stochastic subgra-dient oracles1 of the corresponding local cost function Fi.

Let ðVi;F i;PiÞ, for i ¼ 1; . . . ; N , denote the probabilityspaces, describing the uncertainty associated with individ-ual agents. Here, Vi is the outcome space, F i is a suitables-algebra over Vi, and Pi is the probability distribution overVi. Furthermore, let ðV;F ;PÞ be the joint probability spaceover those spaces. After agent-i’s call at instant t, for anygiven point wwi 2 W, the gradient oracle independently drawsa sample vt;i 2 Vi according to the distribution Pi, and pro-duces a vector ggt;iðvt;iÞ such that

EPifggt;iðviÞg 2 @FiðwwiÞ;

where @FiðwwiÞ denotes the sub-differential set of Fi at wwi. Fornotational simplicity, henceforth, we denote the expectationtaken with respect to the probability distribution Pi by Eif�ginstead of EPif�g. Correspondingly, we use Ef�g insteadof EPf�g.

Although the aim of each agent is to minimize F over Wrather than the local cost Fi, the agents can only call local

subgradient oracles ggi;t, for 1 � t � T . In particular, otherlocal cost functions are totally unknown. Therefore, theagents exchange information with each other within theneighborhoods to mitigate the access restriction.

2.3 Special Cases

We note that this problem formulation is general enough,covering for example the following scenario as a specialcase. Consider that the local cost functions are given byFiðwwiÞ ¼ Eiffiðvi;wwiÞg, where fiðvi;wwiÞ is a certain localloss function, which is a strongly convex function of wwi forany fix vi 2 Vi. At each instant t, a new sample vt;i is drawnfrom Vi independently according to the distribution Pi,and agent-i has access to a corresponding subgradient offiðvt;i;wwiÞ at wwi ¼ wwt;i, i.e.,

ggi;tðvt;iÞ 2 @fiðvt;i;wwt;iÞ:

As an example, the local loss function could be as follows:

fiðvi;wwt;iÞ ¼ ‘ðvi;wwt;iÞ þ �

2wwt;i

�� ���� ��2; (4)

where ‘ðvi;wwt;iÞ is a Lipschitz-continuous convex loss func-tion with respect to the second variable wwt;i, which has beenextensively studied in the literature [23], [32], [33], [34] as a�-strongly convex loss function involving regularity terms.2

Here, the aim of each agent is to minimize the sum of theexpected losses (where the expectations are taken over therandom variables vi’s) over the convex set W. To continuewith our example in (4), each agent seeks to minimize

XNi¼1

Eiffiðvi;wwt;iÞg ¼XNi¼1

Eif‘ðvi;wwt;iÞg þ �

2wwt;i

�� ���� ��2: (5)

We emphasize that the formulation in (5) covers a widerange of practical loss functions. As an example, fordi : Vi ! R and uui : Vi ! Rm, when ‘ðvi;wwt;iÞ ¼ ðdiðviÞ�wwT

t;iuuiðviÞÞ2, we consider the regularized squared error loss;and when ‘ðvi;wwt;iÞ ¼ maxf0; 1� diðviÞwwT

t;iuuiðviÞg, we con-sider the hinge loss. Since we make no assumptions on theloss function fiðvi;wwt;iÞ other than strong convexity, one canalso use different loss functions with their correspondingsubgradients and our results would still hold.

3 MAIN RESULTS

In this section, we present the main results of the paper,where we introduce an algorithm based on the SSD updates,

which leads to a rate of convergence bounded from above

by O NffiffiffiN

pð1�sÞT� �

after T iterates, where N is the number of

agents (nodes). In particular, the rate of convergence of thealgorithm for agent-i is given by

EfF ð �w�wiÞg � minww2W

F ðwwÞ � ON

ffiffiffiffiffiN

p

ð1� sÞT

!;

where �w�wi is the minimizer produced by agent-i, and theexpectation is taken over the randomness of the subgradient

Fig. 1. The neighborhood of agent-i over the distributed network.

1. The agents have a limited budget to call the gradient oracle.

2. Note that � in the regularization term is the same with � in thestrong convexity definition (2). In particular, the regularization termensures that fi is �-strongly convex even when ‘ is not strongly convex.

250 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

Page 4: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

oracles, i.e., with respect to the joint distribution P. In orderto achieve this performance, the proposed method usestime dependent weighted averages of the SSD updates ateach agent together with the adapt-then-combine diffusionstrategy [11]. However, as later explained in this section(See Remark 1), our algorithm can be extended to cover con-sensus in a straightforward manner.

At each time instant t, each agent i has a pre-computedpseudo-solution of problem (3), denoted by wwt;i. With thispseudo-solution agent-i calls the local subgradient oracleand receives ggt;i. Then, agent-i computes the iterate fftþ1;i byprojecting the SSD update of wwt;i ontoW as follows:

fftþ1;i ¼ PW wwt;i � mt;iggt;i� �

;

where mt;i > 0 is a step size. In order to mitigate the accessrestriction to the other oracles, agent-i exchanges fftþ1;i withhis/her neighbors and computes

wwtþ1;i ¼XNj¼1

HHjifftþ1;j; (6)

where HH 2 RNN is the communication matrix of the graphsuch that HHji’s are the combination weights in (6), and theweight HHji for any i; j is nonzero if, and only if, i and jare neighbors. We assume that HH is an irreducible and aperiodic doubly stochastic matrix, i.e., HHi;j � 0 8i; j andHH1 ¼ HHT1 ¼ 1. We emphasize that this assumption is notrestrictive, and previous analyses in the literature also makesimilar assumptions [6], [7], [8], [9]. Furthermore, theassumption holds for many communication strategies suchas the Metropolis rule [3]. At each instant, agent-i also com-putes a time-variant weighted average as follows:

�w�wtþ1;i ¼ t

tþ 2�w�wt;i þ 2

tþ 2wwtþ1;i: (7)

After consuming the budget to call subgradient oracles, i.e.,after T calls, agent-i has �w�wi ¼ �w�wTþ1;i as the minimizer of F .The complete description of the algorithm can be found inAlgorithm 1.

Algorithm 1. Time Variable Weighting (TVW)

1: Initialize �w�w1;i ¼ ww1;i 2 W, 8i, arbitrarily.2: for t ¼ 1 to T do3: for i ¼ 1 toN do4: Call the subgradient oracle to obtain ggt;i for wwt;i.5: cctþ1;i ¼ wwt;i � mt;iggt;i % SSD update6: fftþ1;i ¼ PWðcctþ1;iÞ % Projection7: Exchange fftþ1;i with the neighbors.

8: wwtþ1;i ¼PN

j¼1 HHjifftþ1;j %Diffusion

9: �w�wtþ1;i ¼ ttþ2 �w�wt;i þ 2

tþ2wwtþ1;i %Weighting

10: if t ¼ T then11: �w�wi ¼ �w�wtþ1;i % Solution12: end if13: end for14: end for

To achieve the aforementioned result, we first introducethe following lemma, which provides an upper bound onthe performance of the average parameter vector.

Lemma 1. Assume that for any given ww 2 W, the expectedsquared norm of any produced subgradient oracle is bounded byG2, i.e., Ei ggij jj j2� G2 8i and mt;i ¼ mt. Let

wwt :¼ 1

N

XNi¼1

wwt;i and ww :¼ argminww2W

F ðwwÞ: (8)

Then, Algorithm 1 yields3

E wwtþ1 � wwj jj j2�ð1� �mtÞE wwt � wwj jj j2

� 2mt

N

�F ðwwÞ � EfF ðwwtÞg

þ 4G2m2t

þ 2mtG

N

XNi¼1

2

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2qþ

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � cctþ1;i

�� ���� ��2q �:

This lemma provides an upper bound on the rate of con-vergence and the squared deviation of the average parame-ter vector wwt. It provides an intermediate step to relate theperformance of the parameter vector at each agent to thebest parameter vector. We point out that the assumption inLemma 1 is practically a boundedness condition that iswidely used to analyze the performance of SSD based algo-rithms [33], [34]. We emphasize that our algorithm does notneed to know this upper bound and it is only used in ourtheoretical derivations.

Proof. In order to efficiently manage the recursions, we firstconsider the projection operation and let

xxt;i :¼ PWðcctþ1;iÞ � cctþ1;i: (9)

Then, we can compactly represent the averaged estima-tion parameter wwt (defined in (8)) in a recursive manneras follows [6]

wwtþ1 ¼ 1

N

XNj¼1

XNi¼1

HHij wwt;i � mtggt;i þ xxt;i

� �" #

¼ wwt þ 1

N

XNi¼1

xxt;i � mtggt;i� �

;

(10)

where the last line follows since HH is doubly stochastic,i.e.,HH1 ¼ 1.

Hence, the squared deviation of these average iterateswith respect to any ww can be obtained as follows:

wwtþ1 � wwj jj j2 ¼ wwt � ww þ 1

N

XNi¼1

xxt;i � mtggt;i� ������

����������

�����2

¼ wwt � wwj jj j2þ 1

N2

XNi¼1

ðxxt;i � mtggt;iÞ�����

����������

�����2

þ 2

N

XNi¼1

ðxxt;i � mtggt;iÞT ðwwt � wwÞ:

(11)

We first upper bound the second term on the righthand side (RHS) of (11) through triangle inequality asfollows:

3. Due to SSD update and information exchange, the parameters wwt;i

and cct;i, for t ¼ 1; . . . ; T and i ¼ 1; . . . ; N , are F -measurable. Therefore,the expectation is taken with respect to the distribution P.

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 251

Page 5: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

1

N2

XNi¼1

ðxxt;i � mtggt;iÞ�����

����������

�����2

� 1

N2

XNi¼1

xxt;i

�� ���� ��þ mt ggt;i�� ���� �� !2

:

(12)

We then note that

xxt;i

�� ���� �� ¼ PWðcctþ1;iÞ � cctþ1;i

�� ���� ��� wwt;i � cctþ1;i

�� ���� ��¼ mt ggt;i

�� ���� ��;(13)

where the second line follows from the definition ofthe projection operator (1). Thus, we can rewrite (12)as follows:

1

N2

XNi¼1

ðxxt;i � mtggt;iÞ�����

����������

�����2

� 4m2t

N2

XNi¼1

ggt;i�� ���� �� !2

and taking the expectation of both side with respect to P,we obtain

1

N2EXNi¼1

ðxxt;i � mtggt;iÞ�����

����������

�����2

� 4G2m2t ; (14)

since E ggt;i�� ���� �� ggt;j

�� ���� ��� �ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE ggt;i�� ���� ��2q ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

E ggt;j�� ���� ��2q

� G2 forany i 6¼ j due to Cauchy-Schwarz inequality and theboundedness assumption.

We next turn our attention to ½�ggTt;iðwwt � wwÞ� term in(11) and upper bound this term as follows:

� ggTt;iðwwt � wwÞ ¼ �ggTt;iðwwt � wwt;i þ wwt;i � wwÞ� �ggTt;iðwwt � wwt;iÞ þ FiðwwÞ � Fiðwwt;iÞ

� �

2ww � wwt;i

�� ���� ��2(15)

� �ggTt;iðwwt � wwt;iÞ þ �g�gTt;iðwwt � wwt;iÞ þ FiðwwÞ

� FiðwwtÞ � �

2ww � wwt;i

�� ���� ��2��

2wwt;i � wwt

�� ���� ��2 (16)

� FiðwwÞ � FiðwwtÞ þ �g�gt;i�� ���� ��þ ggt;i

�� ���� ��� �wwt � wwt;i

�� ���� ��� �

2ww � wwt;i

�� ���� ��2��

2wwt;i � wwt

�� ���� ��2; (17)

where �g�gt;i 2 @fiðwwtÞ, (15) follows from the �-strong con-vexity of Fi at wwt;i, i.e.,

FiðwwÞ � Fiðwwt;iÞ þ ggTt;iðww � wwt;iÞ þ �

2ww � wwt;i

�� ���� ��2;(16) also follows from the �-strong convexity of Fi at

wwt, i.e.,

Fiðwwt;iÞ � FiðwwtÞ þ �g�gTt;iðwwt;i � wwtÞ þ �

2wwt;i � wwt

�� ���� ��2;and (17) follows from the Cauchy-Schwarz inequality.Summing (17) from i ¼ 1 to N and taking expectation ofboth sides, we obtain

� EXNi¼1

ggTt;iðwwt � wwÞ( )

� 2GXNi¼1

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2qþ F ðwwÞ � EfF ðwwtÞg

� �N

2

XNi¼1

1

NE ww � wwt;i

�� ���� ��2þ wwt;i � wwt

�� ���� ��2n o

� F ðwwÞ � EfF ðwwtÞg þ 2GXNi¼1

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2q

� �N

2E wwt � wwj jj j2;

(18)

where the first inequality follows from the Cauchy-Schwarz inequality and the boundedness assumption,and the last inequality follows from the Jensen’s inequal-ity due to the convexity of the norm operator.

We finally turn our attention to the xxTt;iðwwt � wwÞ term

in (11) and write it as follows:

xxTt;iðwwt � wwÞ ¼ xxTt;iðwwt � cctþ1;iÞ þ xxT

t;iðcctþ1;i � wwÞ� xxT

t;iðwwt � cctþ1;iÞ;

since

xxTt;iðcctþ1;i � wwÞ � �xxT

t;ixxt;i þ ðcctþ1;i �PWðcctþ1;iÞÞT ðww �PWðcctþ1;iÞÞ

� 0;

where ðcctþ1;i �PWðcctþ1;iÞÞT ðww �PWðcctþ1;iÞÞ � 0 due tothe Euclidean projection onto the convex set W [35]. Tak-ing the expectation of both sides, we can upper boundthis term as follows:

E xxTt;iðwwt � wwÞ � E xxt;i

�� ���� �� wwt � cctþ1;i

�� ���� ��� � Gmt

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � cctþ1;i

�� ���� ��2q (19)

by first using the Cauchy-Schwarz inequality, and then

using (13) and the bound E ggt;i�� ���� �� � ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

E ggi;t�� ���� ��2q

� G.

Putting (14), (18), and (19) back in (11), we obtain

E wwtþ1 � wwj jj j2�ð1� �mtÞE wwt � wwj jj j2

� 2mt

N

hF ðwwÞ � EfF ðwwtÞg þG

XNi¼1

�2

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2q

þffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � cctþ1;i

�� ���� ��2q �iþ 4G2m2t :

(20)

This concludes the proof of Lemma 1. tuHaving obtained an upper bound on the performance of

the average parameter vector, we then consider the meansquare deviation of the parameter vectors at each agentfrom the average parameter vector. This lemma will then beused to relate the performance of each individual agent tothe performance of the fully connected distributed system.

Lemma 2. In addition to the assumptions in Lemma 1, assumethat the initial weights at each agent are identically initialized

252 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

Page 6: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

to avoid any bias,4 i.e., ww1;i ¼ ww1;j, 8i; j. Then Algorithm 1yields ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

E wwt � wwt;i

�� ���� ��2q� 2G

ffiffiffiffiffiN

p Xt�1

z¼1

mt�zsz; (21)

and ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � cctþ1;i

�� ���� ��2q� Gmt þ 2G

ffiffiffiffiffiN

p Xt�1

z¼1

mt�zsz; (22)

where 0 � s < 1 is the second largest singular value of thematrixHH.

Proof.We first let

WWt :¼ ½wwt;1; . . . ; wwt;N �; GGt :¼ ½ggt;1; . . . ; ggt;N �; and

XXt :¼ ½xxt;1; . . . ; xxt;N �:Then, we obtain the recursion onWWt as follows:

WWt ¼ WW 1HHt�1 þ

Xt�1

z¼1

XXt�z � mt�zGGt�zð ÞHHz: (23)

Letting eei denote the basis function for the ith dimension,i.e., only the ith entry of eei is 1 whereas the rest are 0, wehave

wwt � wwt;i

�� ���� �� ¼ WWt1

N1� eei

���������

��������

�Xt�1

z¼1

ðXXt�z � mt�zGGt�zÞ 1

N1�HHzeei

���������

��������

þ ww1 � ww1;i

�� ���� ��¼Xt�1

z¼1

ðXXt�z � mt�zGGt�zÞ 1

N1�HHzeei

���������

��������

(24)

�Xt�1

z¼1

XXt�z � mt�zGGt�zj jj jF1

N1�HHzeei

��������

��������;

�Xt�1

z¼1

ð XXt�zj jj jFþmt�z GGt�zj jj jF Þ1

N1�HHzeei

��������

��������

� 2Xt�1

z¼1

mt�z GGt�zj jj jF1

N1�HHzeei

��������

��������

(25)

where (24) follows due to the unbiased initializationassumption, i.e.,

ww1 ¼ ww1;i ¼ ww1;j; 8i; j 2 f1; . . . ; Ngand (25) follows from (13).

We first consider the term 1N 1�HHzeei�� ���� �� of (25) and

define the matrix BB :¼ 1N 11T . Then, we can write

1

N1�HHzeei

��������

�������� ¼ BBeei �HHzeeij jj j

¼ ðBB�HHÞzeeij jj j;(26)

where the last line follows since BBz ¼ BB, 8z � 1.

Now, let

s1ðAAÞ � s2ðAAÞ � � � � � sNðAAÞdenote the singular values of a matrix AA. Then, we canupper bound (26) as follows:

ðBB�HHÞzeeij jj j � s1ðBB�HHÞ ðBB�HHÞz�1eei�� ���� ��;

8z � 1. Therefore, using the above recursion z times to(26), we obtain

ðBB�HHÞzeeij jj j � sz1ðBB�HHÞ eeij jj j

¼ sz1ðBB�HHÞ: (27)

We note that HH and BB are doubly stochastic matrices;and BB is a rank-1 matrix. Let �1ðAAÞ � �2ðAAÞ � � � � ��NðAAÞ denote the eigenvalues of a symmetric matrix AAand LðAÞ :¼ f�1ðAAÞ; . . . ; �NðAAÞg. Since BB is a rank-1matrix, we have �1ðBBÞ ¼ 1 and �kðBBÞ ¼ 0 for k > 1[36]. We want to compute the largest singular valueof BB�HH, yet BB�HH is not a symmetric matrix. There-fore, we check the eigen-spectrum of ðBB�HHÞTðBB�HHÞ ¼ HHTHH �BB, and the matrices HHTHH and BBare commuting. This yields

LðHHTHH �BBÞ � f�1 � �2 : �1 2 LðHHTHHÞ; �2 2 LðBBÞg:Furthermore, ðBB�HHÞT ðBB�HHÞ1 ¼ ðHHTHH �BBÞ1 ¼ 0, whichimplies that the eigen-spectrum of ðBB�HHÞT ðBB� HHÞ isequal to the eigen-spectrum of HHTHH, except the largesteigenvalue of HHTHH, i.e., �1ðHHTHHÞ ¼ 1. Instead of thateigenvalue, the eigen-spectrum of ðBB�HHÞT ðBB�HHÞincludes 0. Thus, we have

s1ðBB�HHÞ ¼ s2ðHHÞ; (28)

and combining (26), (27), and (28), we obtain

1

N1�HHzeei

��������

�������� � sz

2ðHHÞ: (29)

From here on, we denote s :¼ s2ðHHÞ for notational sim-plicity. We also note that 0 � s < 1 sinceHH is an irreduc-ible and aperiodic doubly stochastic matrix [30], [37].

Using (29) in (25), we obtain

wwt � wwt;i

�� ���� �� � 2Xt�1

z¼1

mt�zsz GGt�zj jj jF : (30)

Taking the expectation of both sides and noting that

E GGt�zj jj j2F ¼ EXNi¼1

ggt�z;i

�� ���� ��2( )

� G2N;

we can rewrite (30) as follows:ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2q� 2G

ffiffiffiffiffiN

p Xt�1

z¼1

mt�zsz: (31)

An upper bound for the term wwt � cctþ1;i

�� ���� �� can beobtained as

4. This is basically an unbiasedness condition, which is reasonablesince the objective weight ww is completely unknown to us. Eventhough the initial weights are not identical, our analyses still hold,albeit with small additional excess terms.

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 253

Page 7: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

wwt � cctþ1;i

�� ���� �� ¼ wwt � wwt;i þ mtggt;i�� ���� ��

� wwt � wwt;i

�� ���� ��þ mt ggt;i�� ���� ��;

where the last line follows from the triangle inequality.Taking square and then expectation of both sides, weobtain the following upper bound

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � cctþ1;i

�� ���� ��2q� Gmt þ 2G

ffiffiffiffiffiN

p Xt�1

z¼1

mt�zsz: (32)

This concludes the proof of Lemma 2. tuThe results in Lemmas 1 and 2 are combined in the fol-

lowing theorem to obtain a regret bound on the perfor-mance of the proposed algorithm. This theorem illustratesthe convergence rate of our algorithm (i.e., Algorithm 1)

over distributed networks. The upper bound on the regret

O NffiffiffiN

pð1�sÞT� �

follows since each agent can only have access to

the other subgradient oracles through the exchange of infor-mation among the agents. Reference [22] provides a lowerbound on the rate of convergence of any algorithm to mini-mize a Lipschitz and strongly convex function for a singleagent system with T oracle calls as O 1

T

� �. Over a central-

ized network, the lower bound becomes O 1NT

� �since at

each time instant, the centralized processor has access toN oracles instead of 1. Therefore, the upper bound on

the rate of convergence, i.e., O NffiffiffiN

pð1�sÞT� �

, matches the lower

bounds presented in [22] up to constant terms,5 hencethe shown dependency of the convergence rate of thealgorithm on T is optimal.

The computational complexity of the algorithm intro-duced is on the order of the computational complexity ofthe SSD iterates up to constant terms. Furthermore, the com-munication load of the proposed method is the same as thecommunication load of the SSD algorithm. On the otherhand, by using a time-dependent averaging of the SSD iter-ates, our algorithm achieves a substantially improved per-formance as shown in Theorem 1 and illustrated throughour simulations in Section 4.

Theorem 1. Under the assumptions in Lemmas 1 and 2, Algo-rithm 1 with learning rate mt ¼ 2

�ðtþ1Þ and weighted parameters�w�wt;i achieves the following guaranteed convergence rate

E F �w�wTþ1;i

� �� � F ðwwÞ � 4NG2

�ðT þ 2Þ 3þ 8sffiffiffiffiffiN

p

1� s

!; (33)

for all T � 1, where 0 � s < 1 is the second largest singularvalue of the matrixHH.

This theorem says that although the agents use only localgradient oracle calls to train their parameter vectors, theyasymptotically achieve the performance of the centralizedprocessor because of the information diffusion over the net-work. This result shows that each agent acquires the infor-mation contained in the gradient oracles at every otheragent and suffers no regret asymptotically as the number ofgradient oracle calls at each agent increases.

Proof. According to Lemmas 1 and 2, we have

E F ðwwtÞf g � F ðwwÞ� N

2mt

E ð1� �mtÞ wwt � wwj jj j2� wwtþ1 � wwj jj j2n o

þ 3NG2mt þ 6NffiffiffiffiffiN

pG2Xt�1

z¼1

mt�zsz:

(34)

From the convexity of the cost functions, we also have

E FiðwwtÞ � Fiðwwt;jÞ� � E ggTt;i;jðwwt � wwt;jÞ

n o; (35)

8i; j 2 f1; . . . ; Ng, where

ggt;i;j 2 @Fiðwwt;jÞ:

Here, we can rewrite (35) as follows:

E Fiðwwt;jÞ � FiðwwtÞ� � E ggTt;i;jðwwt;j � wwtÞ

n o� E ggt;i;j

�� ���� �� wwt;j � wwt

�� ���� ��� � G

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt;j � wwt

�� ���� ��2q;

(36)

where the second line follows from Cauchy Schwarzinequality and the last line follows from the bounded-ness assumption. Summing (36) from i ¼ 1 to N , weobtain

E F ðwwt;jÞ � F ðwwtÞ� � NG

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt;j � wwt

�� ���� ��2q: (37)

Using Lemma 2 in (37), we get

E F ðwwt;jÞ � F ðwwtÞ� � 2N

ffiffiffiffiffiN

pG2Xt�1

z¼1

mt�zsz: (38)

We then add (34) and (38) to obtain

E F ðwwt;jÞ� � F ðwwÞ� N

2mt

E ð1� �mtÞ wwt � wwj jj j2� wwtþ1 � wwj jj j2n o

þ 3NG2mt þ 8NffiffiffiffiffiN

pG2Xt�1

z¼1

mt�zsz:

(39)

Multiplying both sides of (39) by t and summing fromt ¼ 1 to T yields [34]

XTt¼1

t E F ðwwt;jÞ� � F ðwwÞ�

� Nð1� �m1Þ2m1

E ww1 � wwj jj j2� TN

2mT

E wwTþ1 � wwj jj j2

þXTt¼2

N

2

tð1� �mtÞmt

� t� 1

mt�1

�E wwt � wwj jj j2

þ 3NG2XTt¼1

tmt þ 8NffiffiffiffiffiN

pG2XTt¼1

tXt�1

z¼1

mt�zsz:

(40)

5. The number of agents, N , is fixed and independent of the budgetto call the oracles, i.e., T .

254 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

Page 8: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

Next, we observe that

XTt¼1

Xt�1

z¼1

tmt�zsz ¼

XTt¼1

XTz¼1

tmt�zszIft > zg (41)

¼XTz¼1

szXTt¼zþ1

tmt�z

�XTz¼1

szXTt¼1

tmt

� s

1þ s

XTt¼1

tmt;

(42)

where Ift> zg is the indicator function and (42) followssince 0 � s < 1. Using (42) in (40) and insertingmt ¼ 2

�ðtþ1Þ, we obtain

XTt¼1

t EfF ðwwt;jÞg � F ðwwÞ�

� ��NT ðT þ 1Þ4

E wwTþ1 � wwj jj j2

þ 3NG2 þ 8NffiffiffiffiffiN

pG2 s

1� s

� �XTt¼1

2t

�ðtþ 1Þ

� ��NT ðT þ 1Þ4

E wwTþ1 � wwj jj j2

þ 2NG2T

�3þ 8

ffiffiffiffiffiN

p s

1� s

� �;

(43)

where the last line follows since ttþ1 � 1. Dividing both

sides of (43) byPT

t¼1 t ¼ T ðTþ1Þ2 , we obtain

E2

T ðT þ 1ÞXTt¼1

t F ðwwt;jÞ � F ðwwÞ� ( )

� ��N

2E wwTþ1 � wwj jj j2

þ 4NG2

�ðT þ 1Þ 3þ 8ffiffiffiffiffiN

p s

1� s

� �:

(44)

Since Fi’s are convex for all i 2 f1; . . . ; Ng, F is also con-vex. Thus, from Jensen’s inequality, we can write

E F2

T ðT þ 1ÞXTt¼1

t wwt;j

!( )� F ðwwÞ

� E2

T ðT þ 1ÞXTt¼1

t F ðwwt;jÞ � F ðwwÞ� �( ):

(45)

Combining (44) and (45), we obtain

E F2

T ðT þ 1ÞXTt¼1

t wwt;j

!( )� F ðwwÞ

� ��N

2E wwTþ1 � wwj jj j2

þ 4NG2

�ðT þ 1Þ 3þ 8ffiffiffiffiffiN

p s

1� s

� �:

(46)

Note that the weighting step in Algorithm 1, i.e., (7),leads to

wwT;j ¼

=T � 1

T þ 1

=T � 2

T

=T � 3=T � 1� � �

=2 =2þ 2

1 =1þ 2

2 =0þ 2ww1;j

þ

=T � 1

T þ 1

=T � 2

T

=T � 3=T � 1� � �

=3 =3þ 2

2 =2þ 2

2 =1þ 2ww2;j

þ � � � þ 2

T þ 1wwT;j

¼ 2

T ðT þ 1ÞXTt¼1

t wwt;j:

(47)

This concludes the proof of Theorem 1. tuHence, using the weighted average �w�wt;i instead of the

original SSD iterates wwt;i, we can achieve a convergence rate

of O NffiffiffiN

pð1�sÞT� �

. The denominator T of this regret bound fol-

lows since we use a time-varian‘t weighting of the SSD iter-ates. The linear dependency to the network size followssince we add N different cost functions, i.e., one corre-sponding to each agent. Finally, the sub-linear dependencyto the network size results from the diffusion of the parame-ter vector over the distributed network.

We note that the upper bound in (33) includes s, whichcan depend on the network sizeN . In particular, for differentcommunication matrices, the corresponding upper boundson the second largest singular value can be included in (33).As an example, Reference [38] shows that the second largestsingular value of the lazymetropolis matrix, defined by

HHij ¼1

max ni;njf g ; if j 2 N i n i0; if j =2 N i12 � 1

2

Pj2N ini HHij; if i ¼ j;

8><>: (48)

is bounded from above by 1� 171N2, which implies s

1�s<

OðN2Þ.In the following corollary, we provide an MSD guarantee

on the weighted parameters �w�wt;i.

Corollary 1. Under the assumptions in Lemmas 1 and 2, Algo-rithm 1 with learning rate mt ¼ 2

�ðtþ1Þ and weighted parameters�w�wt;i guarantees the following MSD

E �w�wTþ1;i � ww�� ���� ��2� 8NG2

�2ðT þ 2Þ 3þ 8sffiffiffiffiffiN

p

1� s

!;

for all T � 1, where 0 � s < 1 is the second largest singularvalue of the matrixHH.

Proof. This follows from Theorem 1 (33) and �-strong con-vexity (2) of F at ww since 0 2 @F ðwwÞ. tuIn the following corollary, we then consider the perfor-

mance of the average SSD iterate instead of the time-variantweighted iterate in (47). Note that even though the agentshave consumed their budgets to call the subgradient oracle,they can continue to exchange information, which averagesout the iterates. We show that the average SSD iterate

achieves an MSD of OffiffiffiN

pð1�sÞT� �

. This MSD follows due to the

number of gradient oracle calls and diffusion regret overthe distributed network.

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 255

Page 9: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

Corollary 2. Under the assumptions of Lemmas 1 and 2,Algorithm 1 with learning rate mt ¼ 2

�ðtþ1Þ yields the followingguaranteed MSD

E wwTþ1 � wwj jj j2� 8G2

�2ðT þ 1Þ 3þ 8sffiffiffiffiffiN

p

1� s

!: (49)

for all T � 1, where wwt ¼ 1N

PNi¼1 wwt;i and 0 � s < 1 is the

second largest singular value of the matrixHH.

Proof. This follows from (46) and (47) since EfF ð �w�wT;jÞg�F ðwwÞ � 0. tu

Remark 1. Algorithm 1 can be generalized to apply to con-sensus in a straightforward manner, while the perfor-mance guarantee in Theorem 1 still holds up to constant

terms, i.e., we still have a convergence rate of O NffiffiffiN

pð1�sÞT� �

.

For the consensus strategy, the lines 5–8 of Algorithm 1

would be replaced by the following update

wwtþ1;i ¼ PWXNj¼1

HHjiwwt;j � mtggt;i

!: (50)

Hence, we have the following recursion on the parameter

vectors

WWt ¼ WW 1HHt�1 �

Xt�1

z¼1

XXt�z � mt�zGGt�zð ÞHHz�1;

instead of the one in (23). Under this modification,

Lemma 2 can be updated as follows:

ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiE wwt � wwt;i

�� ���� ��2q� 2G

ffiffiffiffiffiN

p Xt�1

z¼1

mt�zsz�1: (51)

This loosens the upper bounds in (25) and (32) by a factorof 1=s (note that 0 � s < 1). Therefore, diffusion strate-

gies achieves a better convergence performance com-

pared to the consensus strategy.

We note that the proposed algorithm leads to the theoret-ical bounds on the convergence rate for a certain step size,which is mt ¼ 2

�ðtþ1Þ. Otherwise, the algorithm can also beused with different step sizes yet not necessarily deliveringsuch theoretical performance guarantees. Furthermore,even though all the agents use mt ¼ 2

�ðtþ1Þ, the only necessaryinformation for them to keep how many times they havecalled the subgradient oracles and then they can all use thesame step size 2=� and scale it by 1=ðtþ 1Þ. We considerthat the cost function F is �-strongly convex, i.e., the agentshave the knowledge of � even though they do not knowwhat the function is.

4 SIMULATIONS

In this section, we first examine the performance of the pro-posed algorithms for various distributed network topolo-gies, namely the star, the circle, and a random networktopologies (which are shown in Fig. 3). In all cases, we havea network of N ¼ 20 agents where each agent i at time t,observes the data dt;i ¼ wwT

0 uut;i þ vt;i, i ¼ 1; . . . ; N , where theregression vector uut;i and the observation noise vt;i are

generated from i.i.d. zero mean Gaussian processes for allt � 1. The variance of the observation noise is s2

v;i ¼ 0:1 for alli ¼ 1; . . . ;N , whereas the auto-covariance matrix of theregression vector uut;i 2 R5 is randomly chosen for each agenti ¼ 1; . . . ;N such that the signal-to-noise ratio (SNR) over thenetwork varies between �15 dB to 10 dB (see Fig. 2). Theparameter of interest, ww0 2 R5, is randomly chosen from azero mean Gaussian process and normalized to have a unitnorm, i.e., ww0j jj j ¼ 1. We use the well-knownMetropolis com-bination rule [3] to set the combinationweights as follows:

HHij ¼1

max ni;njf g ; if j 2 N i n i0; if j =2 N i

1�Pj2N ini HHij; if i ¼ j

8><>: (52)

where ni is the number of neighboring agents for agent i.For this set of experiments, we consider the squared error

loss, i.e., ‘ðwwt;i;uut;i; dt;iÞ ¼ ðdt;i � wwTt;iuut;iÞ2 as our loss func-

tion. In the figures, CSS represents the distributed constantstep-size SSD algorithm of [11], VSS represents the distrib-uted variable step-size SSD algorithm of [6], UW representsthe distributed version of the uniform weighted SSD algo-rithm of [23], and TVW represents the distributed time vari-ant weighted SSD algorithm introduced in this paper. Thestep-sizes of the CSS-1, CSS-2, and CSS-3 algorithms are setto 0.05, 0.1, and 0.2, respectively, at each agent and thelearning rates of the VSS and UW algorithms are set to1=ð�tÞ as noted in [6], [23], whereas the learning rate of theTVW algorithm is set to 2=ð�ðtþ 1ÞÞ as noted in Theorem 1,where � ¼ 0:01. These learning rates are chosen specificallyto guarantee a fair performance comparison between thesealgorithms according to the corresponding algorithmdescriptions stated in this paper and in [6], [23].

In the left column of Fig. 3, we compare the normalizedtime accumulated error performances of these algorithmsunder different network topologies in terms of the globalnormalized cumulative error (NCE) measure, i.e.,

NCEðtÞ ¼ 1

Nt

XNi¼1

Xtt¼1

ðdt;i � wwTt;iuut;iÞ2:

Fig. 2. SNRs of the agents in the distributed network.

256 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

Page 10: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

Fig. 3. NCE (left column) and MSD (right column) performances of the proposed algorithms under the star (first row), the circle (second row), and arandom (third row) network topologies, under the squared error loss function averaged over 200 trials.

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 257

Page 11: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

Additionally, in the right column of Fig. 3, we compare theperformances of the algorithms in terms of the global MSDmeasure, i.e.,

MSDðtÞ ¼ 1

N

XNi¼1

ww0 � wwt;i

�� ���� ��2: (53)

In the figures, we have plotted the NCE and MSE perform-ances of the proposed algorithms over 200 independent tri-als to avoid any bias.

As can be seen in the Fig. 3, the TVW algorithm substan-tially outperforms its competitors and achieves a muchsmaller error performance. This superior performance ofour algorithm is achieved thanks to the time-dependentweighting of the regression parameters, used to obtain afaster convergence rate with respect to the rest of the algo-rithms. Hence, by using a certain time varying weighting ofthe SSD iterates, we obtain a significantly improved conver-gence performance compared to the state-of-the-artapproaches in the literature. Furthermore, the performanceof our algorithm is robust against the network topology,whereas the competitor algorithms may not provide satis-factory performances under different network topologies.

We next consider the classification tasks over the bench-mark data sets: Covertype6 and quantum.7 For this set ofexperiments, we consider the hinge loss, i.e., ‘ðwwt;i;uut;i; dt;iÞ ¼ maxf0; 1� dt;iww

Tt;iuut;ig2 as our loss function. The

regularization constant is set to � ¼ 1=T , where the stepsizes of the TVW, UW, and VSS algorithms are set as in theprevious experiment. The step sizes of the CSS-1, CSS-2,and CSS-3 algorithms are set to 0.02, 0.05, and 0.1 for thecovertype data set, whereas the step sizes of the CSS-1,CSS-2, and CSS-3 algorithms are set to 0.01, 0.02, and 0.05for the quantum data set. These learning rates are chosen toillustrate the tradeoff between the convergence speed andthe steady state performance of the constant step size SSD

methods. The network sizes are set to N ¼ 20 and N ¼ 50for the covertype and quantum data sets, respectively.

In Figs. 4 and 5, we illustrate the performances of the sixalgorithms for various training data lengths. In particular,we train the parameter vectors at each agent using a certainlength of training data and test the performance of the finalparameter vector over the entire data set. We provide aver-aged results over 250 and 100 independent trials for covert-ype and quantum data sets, respectively, and present themean and variance of the normalized accumulated hingeerrors. These figures illustrate that the proposed TVW algo-rithm significantly outperforms its competitors. Althoughthe performances of the UW and VSS algorithms are compa-rably robust over different iterations, the TVW algorithmprovides a smaller accumulated loss. On the other hand,the variances of the constant step size methods highlydeteriorate as the step size increases. Although decreasingthe step size yields more robust performance for theseconstant step size algorithms, the TVW algorithm

Fig. 4. Normalized accumulated errors of the six algorithms versus train-ing data length for cover type data averaged over 250 trials for a networkof size 20.

Fig. 5. Normalized accumulated errors of the six algorithms versus train-ing data length for quantum data averaged over 100 trials for a networkof size 50.

Fig. 6. Global MSD over the star networks with different sizes:50, 100, 200, and 500.

6. https://www.csie.ntu.edu.tw/�cjlin/libsvmtools/datasets/7. http://osmot.cs.cornell.edu/kddcup/

258 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017

Page 12: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

provides a significantly smaller steady-state cumulativeerror with respect to these methods.

Finally, we note that the upper bounds in Theorem 1 andCorollary 1 directly depend on the number of agents in addi-tion to the second largest singular value of the combinationmatrix, s. In the following numerical examples, we examinehow the MSD of the algorithm scales with the network size.To this end, we consider the setup for Fig. 3b with differentnetwork sizes: 50,100,200, and 500. Fig. 6 shows how the timeevolution of the global MSD scales with increasing networksizes from 50 to 500. Furthermore, Fig. 7 shows how s scaleswith the network size and correspondingly we observethat 1=ð1� sÞ scales with N . We also note that the globalMSD measure (53) is averaged across the network. There-fore, the corresponding upper bound on the global MSD(See Corollary 1) scales with

ffiffiffiffiffiN

p=ð1� sÞ. However, in Fig. 6,

we observe that when the network size scales by 10, e.g.,from 50 to 500, the global MSD scales by 10 dB rather than15dB. This raises the possibility that the dependency of theupper bound on the network size might be tightened furtherand formulating the upper bound, which is also optimal interms of network size complexity, can be an interestingfuture research direction.

5 CONCLUSION

We have studied distributed strongly convex optimizationover distributed networks, where the aim is to minimize asum of unknown convex objective functions. We have intro-duced an algorithm that uses a limited number of gradientoracle calls to these objective functions and achieves an opti-

mal convergence rate of O NffiffiffiN

pð1�sÞT� �

after T gradient updates

at each agent. This level of performance is achieved byusing a certain time-dependent weighting of the SSD iter-ates at each agent. Additionally, the weighted parameters

achieve a guaranteed mean square deviation (MSD) of

O NffiffiffiN

pð1�sÞT� �

after T gradient updates. The computational com-

plexity and the communication load of the proposedapproach is the same as with the state-of-the-art methods inthe literature up to constant terms. We have also provedthat the average SSD iterate, which can be attained if theagents continue to exchange information without gradient

updates, achieves a guaranteed MSD of OffiffiffiN

pð1�sÞT� �

after T

gradient oracle calls. We have illustrated the superior con-vergence rate of our algorithm with respect to the state-of-

the-art methods in the literature. Some future directions ofresearch on this topic include the computation of conver-gence rate bounds for the heterogeneous case, where agentscan have different step sizes and/or learning rates, andasynchronous distributed computation as in [39].

ACKNOWLEDGMENTS

This work is supported in part by TUBITAK Contract No115E917, in part by the U.S. Office of Naval Research (ONR)MURI grant N00014-16-1-2710, and in part by NSF undergrant CCF 11-11342.

REFERENCES

[1] A. Yazicioglu, M. Egerstedt, and J. Shamma, “Formation of robustmulti-agent networks through self-organizing random regulargraphs,” IEEE Trans. Netw. Sci. Eng., vol. 2, no. 4, pp. 139–151,Oct.-Dec. 2015.

[2] D. Mateos-Nunez and J. Cortes, “Distributed online convex opti-mization over jointly connected digraphs,” IEEE Trans. Netw. Sci.Eng., vol. 1, no. 1, pp. 23–37, Jan.-Jun. 2014.

[3] A. H. Sayed, “Adaptive networks,” Proc. IEEE, vol. 102, no. 4,pp. 460–497, Apr. 2014.

[4] A. H. Sayed, S.-Y. Tu, J. Chen, X. Zhao, and Z. J. Towfic,“Diffusion strategies for adaptation and learning over networks:An examination of distributed strategies and network behavior,”IEEE Signal Process. Mag., vol. 30, no. 3, pp. 155–171, May 2013.

[5] J.-J. Xiao, A. Ribeiro, Z.-Q. Luo, and G. B. Giannakis, “Distributedcompression-estimation using wireless sensor networks,” IEEESignal Process. Mag., vol. 23, no. 4, pp. 27–41, Jul. 2006.

[6] F. Yan, S. Sundaram, S. V. N. Vishwanathan, and Y. Qi,“Distributed autonomous online learning: Regrets and intrinsicprivacy-preserving properties,” IEEE Trans. Knowl. Data Eng.,vol. 25, no. 11, pp. 2483–2493, Nov. 2013.

[7] S. Sundhar Ram , A. Nedic, and V. V. Veeravalli, “Distributedstochastic subgradient projection algorithms for convex opti-mization,” J. Optimization Theory Appl., vol. 147, no. 3, pp. 516–545,Dec. 2010.

[8] A. Nedic and A. Ozdaglar, “Distributed subgradient methods formulti-agent optimization,” IEEE Trans. Autom. Control, vol. 54,no. 1, pp. 48–61, Jan. 2009.

[9] I. Lobel and A. Ozdaglar, “Distributed subgradient methods forconvex optimization over random networks,” IEEE Trans. Autom.Control, vol. 56, no. 6, pp. 1291–1306, Jun. 2011.

[10] Z. J. Towfic and A. H. Sayed, “Adaptive penalty-based distributedstochastic convex optimization,” IEEE Trans. Signal Process.,vol. 62, no. 15, pp. 3924–3938, Aug. 2014.

[11] C. G. Lopes and A. H. Sayed, “Diffusion least-mean squares overadaptive networks: Formulation and performance analysis,” IEEETrans. Signal Process., vol. 56, no. 7, pp. 3122–3136, Jul. 2008.

[12] F. S. Cattivelli and A. H. Sayed, “Diffusion LMS strategies for dis-tributed estimation,” IEEE Trans. Signal Process., vol. 58, no. 3,pp. 1035–1048, Mar. 2010.

[13] J. Chen, C. Richard, and A. H. Sayed, “Multitask diffusion adapta-tion over networks,” IEEE Trans. Signal Process., vol. 62, no. 16,pp. 4129–4144, Aug. 2014.

[14] S.-Y. Tu and A. H. Sayed, “Distributed decision-making overadaptive networks,” IEEE Trans. Signal Process., vol. 62, no. 5,pp. 1054–1069, Mar. 2014.

[15] J. Chen, C. Richard, and A. H. Sayed, “Diffusion LMS over multi-task networks,” IEEE Trans. Signal Process., vol. 63, no. 11,pp. 2733–2748, Jun. 2015.

[16] X. Zhao and A. H. Sayed, “Performance limits for distributed esti-mation over LMS adaptive networks,” IEEE Trans. Signal Process.,vol. 60, no. 10, pp. 5107–5124, Oct. 2012.

[17] J. Chen and A. H. Sayed, “Diffusion adaptation strategies for dis-tributed optimization and learning over networks,” IEEE Trans.Signal Process., vol. 60, no. 8, pp. 4289–4305, Aug. 2012.

[18] S.-Y. Tu and A. H. Sayed, “Diffusion strategies outperform consen-sus strategies for distributed estimation over adaptive networks,”IEEE Trans. Signal Process., vol. 60, no. 12, pp. 6217–6234, Dec. 2012.

[19] J. Chen and A. H. Sayed, “Distributed pareto optimization viadiffusion strategies,” IEEE J. Sel. Topics Signal Process., vol. 7, no. 2,pp. 205–220, Apr. 2013.

Fig. 7. Second largest singular value, s, of the combination matrix Hover the star networks with different sizes: 50, 100, 200, and 500.

SAYIN ET AL.: STOCHASTIC SUBGRADIENT ALGORITHMS FOR STRONGLY CONVEX OPTIMIZATION OVER DISTRIBUTED NETWORKS 259

Page 13: 248 IEEE TRANSACTIONS ON NETWORK SCIENCE AND …

[20] X. Zhao and A. H. Sayed, “Distributed clustering and learningover networks,” IEEE Trans. Signal Process., vol. 63, no. 13,pp. 3285–3300, Jul. 2015.

[21] Z. J. Towfic and A. H. Sayed, “Stability and performance limits ofadaptive primal-dual networks,” IEEE Trans. Signal Process.,vol. 63, no. 11, pp. 2888–2903, Jun. 2015.

[22] A. Agarwal, P. L. Bartlett, P. Ravikumar, and M. J. Wainwright,“Information-theoretic lower bounds on the oracle complexity ofstochastic convex optimization,” IEEE Trans. Inf. Theory, vol. 58,no. 5, pp. 3235–3249, May 2012.

[23] A. Rakhlin, O. Shamir, and K. Sridharan, “Making gradientdescent optimal for strongly convex stochastic optimization,” inProc. 29th Int. Conf. Mach. Learn., 2012, pp. 449–456.

[24] K. Slavakis, G. B. Giannakis, and G. Mateos, “Modeling and opti-mization for big data analytics: (Statistical) learning tools for ourera of data deluge,” IEEE Signal Process. Mag., vol. 31, no. 5,pp. 18–31, Sep. 2014.

[25] I. D. Schizas, G. Mateos, and G. B. Giannakis, “Distributed LMSfor consensus-based in-network adaptive processing,” IEEE Trans.Signal Process., vol. 57, no. 6, pp. 2365–2382, Jun. 2009.

[26] G. Mateos, I. Schizas, and G. Giannakis, “Distributed recursiveleast-squares for consensus-based in-network adaptiveestimation,” IEEE Trans. Signal Process., vol. 57, no. 11, pp. 4583–4588, Nov. 2009.

[27] G. Mateos and G. B. Giannakis, “Distributed recursive least-squares: Stability and performance analysis,” IEEE Trans. SignalProcess., vol. 60, no. 7, pp. 3740–3754, Jul. 2012.

[28] G. Mateos, J. A. Bazerque, and G. B. Giannakis, “Distributedsparse linear regression,” IEEE Trans. Signal Process., vol. 58,no. 10, pp. 5262–5276, Oct. 2010.

[29] P. A. Forero, A. Cano, and G. B. Giannakis, “Distributed clusteringusing wireless sensor networks,” IEEE J. Sel. Topics Signal Process.,vol. 5, no. 4, pp. 707–724, Aug. 2011.

[30] J. C. Duchi, A. Agarwal, and M. J. Wainwright, “Dual averagingfor distributed optimization: Convergence analysis and networkscaling,” IEEE Trans. Autom. Control, vol. 57, no. 3, pp. 592–606,Mar. 2012.

[31] K. Tsianos and M. Rabbat, “Distributed strongly convex opti-mization,” in Proc. 50th Annu. Allerton Conf. Commun. Control Com-put., Oct. 2012, pp. 593–600.

[32] E. Hazan, A. Agarwal, and S. Kale, “Logarithmic regret algo-rithms for online convex optimization,” Mach. Learn., vol. 69,no. 2–3, pp. 169–192, Dec. 2007.

[33] E. Hazan and S. Kale, “Beyond the regret minimization barrier:Optimal algorithms for stochastic strongly-convex optimization,”J. Mach. Learn. Res., vol. 15, pp. 2489–2512, Jul. 2014.

[34] S. Lacoste-Julien,M.W. Schmidt, and F. Bach, “A simpler approachto obtaining anOð1=tÞ convergence rate for the projected stochasticsubgradient method,” Dec. 2012. [Online]. Available: http://arxiv.org/pdf/1212.2002v2.pdf

[35] D. G. Luenberger, Optimization by Vector Space Methods. Hoboken,NJ, USA: Wiley, 1969.

[36] R. A. Horn and C. R. Johnson,Matrix Analysis. Cambridge, UnitedKingdom: Cambridge Univ. Press, 1985.

[37] D. Levin, Y. Peres, and E. Wilmer, Markov Chains and MixingTimes. Providence, Rhode Island, USA: American MathematicalSociety, 2008.

[38] A. Olshevsky, “Linear time average consensus on fixed graphsand implications for decentralized optimization and multi-agentcontrol,” arXiv preprint arXiv:1411.4186, 2016.

[39] S. Li and T. Basar, “Asymptotic agreement and convergence ofasynchronous stochastic algorithms,” IEEE Trans. Autom. Control,vol. AC-32, no. 7, pp. 612–618, Jul. 1987.

Muhammed O. Sayin received the BS and MSdegrees in electrical and electronics engineeringfrom Bilkent University, Ankara, Turkey, in 2013and 2015, respectively. He is currently pursuingthe PhD degree in electrical and computer engi-neering from the University of Illinois at Urbana-Champaign (UIUC). His current research inter-ests include signaling games, dynamic gamesand decision theory, strategic decision making,and stochastic optimization.

N. Denizcan Vanli received the BS and MSdegrees in electrical and electronics engineeringfrom Bilkent University, Ankara, Turkey, in 2013and 2015, respectively. He is currently pursuingthe PhD degree in electrical engineering andcomputer science at Massachusetts Institute ofTechnology, Cambridge, MA. His research inter-ests include convex optimization, online learning,and distributed optimization.

Suleyman S. Kozat received the BS degree withfull scholarship and high honors from Bilkent Uni-versity, Turkey and the MS and PhD degrees inelectrical and computer engineering from the Uni-versity of Illinois at Urbana Champaign, Urbana,IL. After graduation, he joined IBMResearch, T. J.Watson Research Lab, Yorktown, New York, as aresearch staff member (and later became a proj-ect leader) in the Pervasive Speech TechnologiesGroup, where he focused on problems related tostatistical signal processing and machine learn-

ing. While doing the PhD degree, he was also working as a researchassociate at Microsoft Research, Redmond, Washington, in the Cryp-tography and Anti-Piracy Group. He holds several patent inventionsdue to his research accomplishments at IBM Research and MicrosoftResearch. He is currently an associate professor at the Electrical andElectronics Engineering Department at Bilkent University. He is theelected president of the IEEE Signal Processing Society, TurkeyChapter. He coauthored more than 100 papers in refereed highimpact journals and conference proceedings and has several patentinventions (currently used in several different Microsoft and IBM prod-ucts such as the MSN and the ViaVoice). He holds many internationaland national awards. Overall, his research interests include cybersecurity, anomaly detection, big data, data intelligence, adaptive fil-tering and machine learning algorithms for signal processing. He is asenior member of the IEEE.

Tamer Basar received the BSEE degree fromRobert College, Istanbul, the MS, MPhil, andPhD degrees from Yale University. He is with theUniversity of Illinois at Urbana-Champaign, wherehe holds the academic positions of Swanlundendowed chair; Center for Advanced Study pro-fessor of Electrical and Computer Engineering;research professor at the Coordinated ScienceLaboratory; and research professor at the Infor-mation Trust Institute. He is also the director ofthe Center for Advanced Study. He is a member

of the US National Academy of Engineering, member of the EuropeanAcademy of Sciences, IFAC (International Federation of Automatic Con-trol) and SIAM (Society for Industrial and Applied Mathematics), and hasserved as president of IEEE CSS (Control Systems Society), ISDG(International Society of Dynamic Games), and AACC (American Auto-matic Control Council). He has received several awards and recogni-tions over the years, including the highest awards of IEEE CSS, IFAC,AACC, and ISDG, the IEEE Control Systems Award, and a number ofinternational honorary doctorates and professorships. He has more than750 publications in systems, control, communications, and dynamicgames, including books on non-cooperative dynamic game theory,robust control, network security, wireless and communication networks,and stochastic networked control. He was the Editor-in-Chief of Automa-tica between 2004 and 2014, and is currently editor of several bookseries. His current research interests include stochastic teams, games,and networks; distributed algorithms; security; and cyber-physical sys-tems. He is a fellow of the IEEE.

" For more information on this or any other computing topic,please visit our Digital Library at www.computer.org/publications/dlib.

260 IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, VOL. 4, NO. 4, OCTOBER-DECEMBER 2017