Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Not converge when larger batch size is used in Stage2 #56

Open
NickChang97 opened this issue Aug 23, 2023 · 2 comments
Open

Not converge when larger batch size is used in Stage2 #56

NickChang97 opened this issue Aug 23, 2023 · 2 comments

Comments

@NickChang97
Copy link

Hi, the default batch size is 1, did you try the larger batch size. In my experiments, when the batch size >= 4, the loss can not be converged to a satisfied results despite tuning various hyper-parameters.

@Doubiiu
Copy link
Owner

Doubiiu commented Aug 23, 2023

Hi, I didn't try that in our experiments. I think this VQ-Stuff is unstable in terms of training, thus a 'proper' combination of these hyper-parameters is a must to make it work.

@yangyifan18
Copy link

I meet the same problem!I trained VAE with batch size 1, but the results seems to be wrong when test with batch size 8. This does not make sense because batch_size should not affect test performance. I think there might be something wrong with the view() operation in VectorQuantizer

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants