Skip to content

LR defined wrt single gpu batchsize or global batch size? #766

Answered by rwightman
yoni-f asked this question in Q&A
Discussion options

You must be logged in to vote

@yoni-f batch size is per GPU right now, so you need to calculate global based on num gpu for converting between example hparams and your own setup, I realize that's not ideal and intend to add a mode that normalizes lr per 256 batch size but will likely end up in the future 'bits' training code

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@yoni-f
Comment options

@rwightman
Comment options

@yoni-f
Comment options

Answer selected by yoni-f
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants