We consider distributed smooth nonconvex unconstrained optimization over networks, modeled as a connected graph. We examine the behavior of distributed gradient-based algorithms near strict saddle points. Specifically, we establish that (i) the renowned Distributed Gradient Descent (DGD) algorithm likely converges to a neighborhood of a Second-order Stationary (SoS) solution; and (ii) the more recent class of distributed algorithms, based on gradient tracking (termed SONATA), likely converges to exact SoS solutions, thus avoiding (strict) saddle points.
Second-order Guarantees of Gradient Algorithms over Networks
Amir Daneshmand,G. Scutari,V. Kungurtsev
Published 2018 in Allerton Conference on Communication, Control, and Computing
ABSTRACT
PUBLICATION RECORD
- Publication year
2018
- Venue
Allerton Conference on Communication, Control, and Computing
- Publication date
2018-10-01
- Fields of study
Mathematics, Computer Science, Engineering
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-25 of 25 references · Page 1 of 1
CITED BY
Showing 1-13 of 13 citing papers · Page 1 of 1