-
Hey, I ran one of your example training runs (EfficientNet-B2 with RandAugment - 80.4 top-1, 95.1 top-5), and I parsed the .csv to a tensorboard and I noticed something a bit weird. The training loss was a lot noiser than the validation loss.
Why are you reducing the loss before logging it? Thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
@yoni-f I want to log the loss across all processes, not just the one logging. I'd expect the training loss to be noisy with higher aug. But also, if that particular training was w/ weight EMA enabled, the val loss that ends up in the log file will be the EMA weight eval loss so it would be much smoother compared to train |
Beta Was this translation helpful? Give feedback.
@yoni-f I want to log the loss across all processes, not just the one logging. I'd expect the training loss to be noisy with higher aug. But also, if that particular training was w/ weight EMA enabled, the val loss that ends up in the log file will be the EMA weight eval loss so it would be much smoother compared to train