2018
DOI: 10.48550/arxiv.1811.06396
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Asynchronous Stochastic Composition Optimization with Variance Reduction

Abstract: Composition optimization has drawn a lot of attention in a wide variety of machine learning domains from risk management to reinforcement learning. Existing methods solving the composition optimization problem often work in a sequential and single-machine manner, which limits their applications in large-scale problems. To address this issue, this paper proposes two asynchronous parallel variance reduced stochastic compositional gradient (AsyVRSC) algorithms that are suitable to handle large-scale data sets. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 6 publications
0
3
0
Order By: Relevance
“…where the first equality follows from (25) and the row stochasticity of W, the first inequality follows from the fact |||W − I n ||| ≤ 2, the second inequality follows from Assumption 1(c), the fact…”
Section: Proof Of Lemmamentioning
confidence: 99%
See 1 more Smart Citation
“…where the first equality follows from (25) and the row stochasticity of W, the first inequality follows from the fact |||W − I n ||| ≤ 2, the second inequality follows from Assumption 1(c), the fact…”
Section: Proof Of Lemmamentioning
confidence: 99%
“…In the past decades, stochastic gradient decent methods have been well studied for solving stochastic compositional optimization problem, such as, two-timescale scheme method [28,29], sing-timescale scheme method [4,12] and variance reduction based method [14,17,25]. More recently, Gao and Huang [10] study distributed stochastic compositional optimization problem over undirected communication networks, where a gossip-based distributed stochastic gradient descent method and its gradient-tracking version are proposed.…”
Section: Introductionmentioning
confidence: 99%
“…When solving either (1.8) or (1.4), most of the existing work is devoted to developing stochastic oraclebased algorithms and their sample complexity analysis for solving these problems. Related work includes two-timescale stochastic approximation algorithms for solving the problem (1.8) (Wang et al, , 2016, variance-reduced algorithms for iteratively solving the SAA counterpart of (1.8) (Lian et al, 2017;Huo et al, 2018;Shen et al, 2018), and a primal-dual functional stochastic approximation algorithm for solving the problem (1.4) (Dai et al, 2017). These methods usually require convexity of the objective in order to obtain an -optimal solution.…”
Section: Related Workmentioning
confidence: 99%