title | abstract | layout | series | publisher | issn | id | month | tex_title | firstpage | lastpage | page | order | cycles | bibtex_author | author | date | address | container-title | volume | genre | issued | extras | ||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Stochastic Approximation with Delayed Updates: Finite-Time Rates under Markovian Sampling |
Motivated by applications in large-scale and multi-agent reinforcement learning, we study the non-asymptotic performance of stochastic approximation (SA) schemes with delayed updates under Markovian sampling. While the effect of delays has been extensively studied for optimization, the manner in which they interact with the underlying Markov process to shape the finite-time performance of SA remains poorly understood. In this context, our first main contribution is to show that under time-varying bounded delays, the delayed SA update rule guarantees exponentially fast convergence of the \emph{last iterate} to a ball around the SA operator’s fixed point. Notably, our bound is \emph{tight} in its dependence on both the maximum delay |
inproceedings |
Proceedings of Machine Learning Research |
PMLR |
2640-3498 |
adibi24a |
0 |
Stochastic Approximation with Delayed Updates: Finite-Time Rates under {M}arkovian Sampling |
2746 |
2754 |
2746-2754 |
2746 |
false |
Adibi, Arman and {D}al Fabbro, Nicol\`{o} and Schenato, Luca and Kulkarni, Sanjeev and Vincent Poor, H. and J. Pappas, George and Hassani, Hamed and Mitra, Aritra |
|
2024-04-18 |
Proceedings of The 27th International Conference on Artificial Intelligence and Statistics |
238 |
inproceedings |
|