Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How are multiple datasets combined during training? #17

Open
DomhnallBoyle opened this issue Jun 9, 2023 · 0 comments
Open

How are multiple datasets combined during training? #17

DomhnallBoyle opened this issue Jun 9, 2023 · 0 comments

Comments

@DomhnallBoyle
Copy link

Hi, thanks for this great work.

I have a question about the section "3.8 Using Additional Training Data" from your paper "Visual Speech Recognition for Multiple Languages in the Wild"

For example, for LRS3 the best WER of 32.1 is achieved by combining the datasets LRW + LRS2 + AVSpeech + LRS3. I was just wondering what way they're combined during training, which of the scenarios would be correct?

Scenario A:

  1. Pretrain using LRW + LRS2 + AVSpeech datasets
  2. Initialise from 1 above, then train on the LRS3 dataset only

Scenario B:

  1. Pretrain using LRW + LRS2 + AVSpeech datasets
  2. Initialise from 1 above, then train on LRW + LRS2 + AVSpeech + LRS3 datasets

Would there be a performance difference between these 2 scenarios?

Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant