-
Hello, I was wondering if in the example training params (https://rwightman.github.io/pytorch-image-models/training_hparam_examples/), are the LR supplied wrt to the batchsize specified (ie 0.016 for batch 128 in the effnetb2 example), or is 0.016 for a global batch size of 256 (128*2 since you are training on 2 gpus)? Hope the question is clear |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
@yoni-f batch size is per GPU right now, so you need to calculate global based on num gpu for converting between example hparams and your own setup, I realize that's not ideal and intend to add a mode that normalizes lr per 256 batch size but will likely end up in the future 'bits' training code |
Beta Was this translation helpful? Give feedback.
@yoni-f batch size is per GPU right now, so you need to calculate global based on num gpu for converting between example hparams and your own setup, I realize that's not ideal and intend to add a mode that normalizes lr per 256 batch size but will likely end up in the future 'bits' training code