You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I used the default setting(which I think is Densenet-12-BC with data augmentation) on cifar-100(via just changing the name of dataset class and the nClasses variable). The training curve looks like this:
Though the training has not ended yet, from training curves for other networks on Cifar-100 I can tell there would be no more major changes in acc. The highest of acc for now is 75.59%, which can only match the reported performance of Densenet-12(depth 40) with data augmentation.
Has any one tested this repo on Cifar-100 yet?
The text was updated successfully, but these errors were encountered:
Hi @wishforgood
I have tried another reimplementation and meet the same problem. The error rate on CIFAR10 with densenet40-nonBC is only 6.0% (5.24% in the repo) but when I test it on Tensorflow it is about 5.4%.
I think it is caused by Pytorch instead of the model
Have you solved that yet?
I used the default setting(which I think is Densenet-12-BC with data augmentation) on cifar-100(via just changing the name of
dataset
class and thenClasses
variable). The training curve looks like this:Though the training has not ended yet, from training curves for other networks on Cifar-100 I can tell there would be no more major changes in acc. The highest of acc for now is 75.59%, which can only match the reported performance of Densenet-12(depth 40) with data augmentation.
Has any one tested this repo on Cifar-100 yet?
The text was updated successfully, but these errors were encountered: