2019
DOI: 10.1007/s00498-019-00249-4
|View full text |Cite
|
Sign up to set email alerts
|

Non-asymptotic error bounds for constant stepsize stochastic approximation for tracking mobile agents

Abstract: This work revisits the constant stepsize stochastic approximation algorithm for tracking a slowly moving target and obtains a bound for the tracking error that is valid for the entire time axis, using the Alekseev non-linear variation of constants formula. It is the first non-asymptptic bound for the entire time axis in the sense that it is not based on the vanishing stepsize limit and associated limit theorems unlike prior works, and captures clearly the dependence on problem parameters and the dimension.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 48 publications
0
2
0
Order By: Relevance
“…Constant stepsize RSAs often converge much faster to a neighborhood of the desired solution. This phenomenon has been observed in off-policy temporal difference learning [52], temporal difference learning with function approximation [33], tracking problems [31], and gradient descent [4], among others. Furthermore, the size of this neighborhood is usually small if the stepsize is small (so too large a stepsize may not be beneficial) [13,7].…”
mentioning
confidence: 83%
“…Constant stepsize RSAs often converge much faster to a neighborhood of the desired solution. This phenomenon has been observed in off-policy temporal difference learning [52], temporal difference learning with function approximation [33], tracking problems [31], and gradient descent [4], among others. Furthermore, the size of this neighborhood is usually small if the stepsize is small (so too large a stepsize may not be beneficial) [13,7].…”
mentioning
confidence: 83%
“…To the best of our knowledge, this paper provides the first finite-time convergence analysis for DSA scheme with biased updates relying on Markov samples. In addition, this work is related to the recent works on non-asymptotic analysis of SA schemes [14][15][16].…”
Section: Introductionmentioning
confidence: 99%