2021
DOI: 10.48550/arxiv.2106.08545
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey on Fault-tolerance in Distributed Optimization and Machine Learning

Shuo Liu

Abstract: The robustness of distributed optimization is an emerging field of study, motivated by various applications of distributed optimization including distributed machine learning, distributed sensing, and swarm robotics. With the rapid expansion of the scale of distributed systems, resilient distributed algorithms for optimization are needed, in order to mitigate system failures, communication issues, or even malicious attacks. This survey investigates the current state of fault-tolerance research in distributed o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 87 publications
0
2
0
Order By: Relevance
“…It is well known that designing robust algorithms to adversarial attacks is nontrivial and challenging for multi-agent consensus, distributed learning and decentralized optimization, etc. (e.g., [92]- [94]). In this respect, adversarial agents were considered recently for DOL in [95], where Byzantine faulty agents can update its variable arbitrarily, which is then transmitted to its neighbors, with the purpose of preventing no-faulty agents from achieving the optimal solution, and individual static regret is ensured by establishing sufficient conditions on the graph topology, the number and location of the adversarial agents.…”
Section: Communication Perspectivementioning
confidence: 99%
“…It is well known that designing robust algorithms to adversarial attacks is nontrivial and challenging for multi-agent consensus, distributed learning and decentralized optimization, etc. (e.g., [92]- [94]). In this respect, adversarial agents were considered recently for DOL in [95], where Byzantine faulty agents can update its variable arbitrarily, which is then transmitted to its neighbors, with the purpose of preventing no-faulty agents from achieving the optimal solution, and individual static regret is ensured by establishing sufficient conditions on the graph topology, the number and location of the adversarial agents.…”
Section: Communication Perspectivementioning
confidence: 99%
“…where H is a set of non-faulty agents. Various methods are proposed to solve Byzantine fault-tolerant optimization or learning problems [34], including robust gradient aggregation [5,11], gradient coding [10], and other methods [52,55].…”
Section: Related Workmentioning
confidence: 99%