-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Vectorizing MNLI inference #5
Comments
and, in the final paper, we can show results using different LMs. BART-MNLI is one and original RoBERTA-MNLI is also one. |
The zero-shot one gives a lot different result than the text-classification task even with the same labels. I will show you on tomorrow's meeting. |
Maybe the reason is because the base model changes from RoBERTa to BART. Maybe we should use a RoBERTa-based model to be fair. https://huggingface.co/roberta-large-mnli Ref: #10 |
so per the discussion this afternoon, we will just vectorize this code below and forget about the zero-shot approach which seems to have issue with long sentences. |
The two segments of code below for MNLI is too slow. Should use vectorized version to speed up.
The approach below computes a pair of sentences each time. It is too slow. Please see whether you can find an API that computes causality between multiple pairs each time.
Huggingface's zero-shot classification task can do it. See my example at the end.
https://github.com/SigmaWe/DocAsRef_0/blob/de4de4b4275e661621bebf3b2f92d8676e2f81c2/mnli/sim.py#L10-L16
https://github.com/SigmaWe/DocAsRef_0/blob/de4de4b4275e661621bebf3b2f92d8676e2f81c2/mnli/eval.py#L22-L26
The text was updated successfully, but these errors were encountered: