2020
DOI: 10.1109/tnnls.2019.2920930
|View full text |Cite
|
Sign up to set email alerts
|

Multidimensional Gains for Stochastic Approximation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 16 publications
(3 citation statements)
references
References 47 publications
0
3
0
Order By: Relevance
“…To determine the optimal set of parameters that fits the training data, the training model has to optimize a loss (objective) function, which penalizes the model when it produces an inaccurate label on a data point. Let l W; x i y i be the loss function for each data sample x i , with W being a matrix (or several matrices) of weights between neurons [18]. Correspondingly, the loss function on a client's local dataset D c is given in Equation (1).…”
Section: Federated Learningmentioning
confidence: 99%
“…To determine the optimal set of parameters that fits the training data, the training model has to optimize a loss (objective) function, which penalizes the model when it produces an inaccurate label on a data point. Let l W; x i y i be the loss function for each data sample x i , with W being a matrix (or several matrices) of weights between neurons [18]. Correspondingly, the loss function on a client's local dataset D c is given in Equation (1).…”
Section: Federated Learningmentioning
confidence: 99%
“…Ref. [6] also assumes direct information on the Hessian is available in a second-order stochastic method, but allows for loss functions more general than the ERF. Ref.…”
Section: Introduction 1problem Contextmentioning
confidence: 99%
“…A floating-point operation is assumed to be either a summation or a multiplication, while transposition requires no FLOPs. For the updating H k step in original 2SPSA, 3p 2 FLOPs are required per (5) and 4p 2 FLOPs are required per (6). In the proposed implementation, 10p FLOPs are required to get ũk and ṽk per (11) and (12), respectively, and 22p 2 /6 + O(p) FLOPs are required to update the symmetric indefinite factorization of H k [24, Thm.…”
mentioning
confidence: 99%