title | abstract | layout | series | publisher | issn | id | month | tex_title | firstpage | lastpage | page | order | cycles | bibtex_author | author | date | address | container-title | volume | genre | issued | extras | |||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Revisiting the Noise Model of Stochastic Gradient Descent |
The effectiveness of stochastic gradient descent (SGD) in neural network optimization is significantly influenced by stochastic gradient noise (SGN). Following the central limit theorem, SGN was initially described as Gaussian, but recently Simsekli et al (2019) demonstrated that the |
inproceedings |
Proceedings of Machine Learning Research |
PMLR |
2640-3498 |
battash24a |
0 |
Revisiting the Noise Model of Stochastic Gradient Descent |
4780 |
4788 |
4780-4788 |
4780 |
false |
Battash, Barak and Wolf, Lior and Lindenbaum, Ofir |
|
2024-04-18 |
Proceedings of The 27th International Conference on Artificial Intelligence and Statistics |
238 |
inproceedings |
|