Replies: 3 comments 6 replies
-
@sowmen I've added the ability to customize the norm layer and activation for a lot of the models in this collection, it shouldn't be too challenging. Actually adding such model defs and training them from scratch isn't currently in my plans. Something (mostly), like below would work for EfficientNet
|
Beta Was this translation helpful? Give feedback.
-
i would like to have this feature as well,,some models with frn layers with tlu could be great for people like us having low gpu vram,,replacing bn with frn+tlu for all different models are not straightforward task,,some of these pretrained models could help people like us a lot but no one working on it for pretraining models :( |
Beta Was this translation helpful? Give feedback.
-
@mobassir94 it's a huge amount of time & work & hardware resources to define and train a series of new models w/ GN so it's unlikely to happen unless someone else takes it on. One of the likely reasons why this isn't common is that GN models are no faster, and actually use more GPU ram than a BN network. You can train w/ stability at smaller batch sizes, but they force you to use smaller batch sizes even sooner than BN based net. |
Beta Was this translation helpful? Give feedback.
-
Group Normalization seems to be effective for low batch training on less powerful h/w. Any plans to replace BN or add options for GN to existing models like EfficientNet? If not can you suggest a way I can implement GN to existing EfficientNet models? I was hoping to try the BiT (Big Transfer) technique.
Beta Was this translation helpful? Give feedback.
All reactions