Back to Search
Start Over
Distributed Gradient Methods with Variable Number of Working Nodes
- Source :
- IEEE Transactions on Signal Processing. 64:4080-4095
- Publication Year :
- 2016
- Publisher :
- Institute of Electrical and Electronics Engineers (IEEE), 2016.
-
Abstract
- We consider distributed optimization where $N$ nodes in a connected network minimize the sum of their local costs subject to a common constraint set. We propose a distributed projected gradient method where each node, at each iteration $k$, performs an update (is active) with probability $p_k$, and stays idle (is inactive) with probability $1-p_k$. Whenever active, each node performs an update by weight-averaging its solution estimate with the estimates of its active neighbors, taking a negative gradient step with respect to its local cost, and performing a projection onto the constraint set; inactive nodes perform no updates. Assuming that nodes' local costs are strongly convex, with Lipschitz continuous gradients, we show that, as long as activation probability $p_k$ grows to one asymptotically, our algorithm converges in the mean square sense (MSS) to the same solution as the standard distributed gradient method, i.e., as if all the nodes were active at all iterations. Moreover, when $p_k$ grows to one linearly, with an appropriately set convergence factor, the algorithm has a linear MSS convergence, with practically the same factor as the standard distributed gradient method. Simulations on both synthetic and real world data sets demonstrate that, when compared with the standard distributed gradient method, the proposed algorithm significantly reduces the overall number of per-node communications and per-node gradient evaluations (computational cost) for the same required accuracy.<br />Comment: submitted to a journal on April 15, 2015; revised on September 23, 2015, and March 10, 2016
- Subjects :
- FOS: Computer and information sciences
Discrete mathematics
0209 industrial biotechnology
Mathematical optimization
Computer Science - Information Theory
Information Theory (cs.IT)
020206 networking & telecommunications
02 engineering and technology
Lipschitz continuity
Set (abstract data type)
Projection (relational algebra)
020901 industrial engineering & automation
Optimization and Control (math.OC)
Signal Processing
Convergence (routing)
FOS: Mathematics
0202 electrical engineering, electronic engineering, information engineering
Node (circuits)
Electrical and Electronic Engineering
Gradient descent
Convex function
Mathematics - Optimization and Control
Gradient method
Mathematics
Subjects
Details
- ISSN :
- 19410476 and 1053587X
- Volume :
- 64
- Database :
- OpenAIRE
- Journal :
- IEEE Transactions on Signal Processing
- Accession number :
- edsair.doi.dedup.....9de79b81e500d3400a37b6de359a4408