-
Notifications
You must be signed in to change notification settings - Fork 340
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tokens in multi-turn setting #30
Comments
If I understand right, there are NO tokens between dialogue sessions. Because one dialogue session is one training example and contains source (utt1 <|eos|> utt2 <|eos|> utt3) and target (utt4). Next session passed to the model as another training sample. |
Thank you liethman, this is very helpful. My current, updated understanding is that the .tsv file must be in the format you described, Then, the prepro.py will create the features, that end with an <|endoftext|> token (id=50256). |
|
I successfully managed to fine-tune the model with input data in this form: A sample training instance is therefore : |
Hi, @ferdinando17 . I am trying to fine-tune the model with my own dataset. I failed to run
Hope to get your reply. Thanks. |
Hi, to ask dialoGPT to predict "i am a admin. " Also, the zeros mean you are not training on all the utterances that follow them, is it what you want? |
Hi, @ferdinando17 , this is what bothers me. In multi-turn dialog, we have several previous turns as context, one user turn as the question and one system turn as the answer. Through your explanation, I realized that it should be
as the example in the training/fine-tuning dataset, where only the first sentence should be 0.0 and the remaining sentences should be 1.0 to train/fine-tune the model regardless of the user turn or the system turn. |
Are you applying it to task-oriented dialogue? I understand that the 0.0 are for those sentences that you want to filter, the authors used it to avoid training on offensive language. I hope it makes sense. |
Hi, @ferdinando17. Thank you for your reply. |
Ok, I see. I disagree, but of course I might be wrong. Let me know if you find evidence of the contrary. |
hi gays,how do i deal with datasets like this : is this correct? |
also i just wonder what should the validation set looks like |
Hi,
thanks for making the work available and for the explanations.
From the paper I understand that a training instance is a dialogue session, made up of several dialogue turns concatenated and ended by the end-of-text token.
Based on this and on what dreasysnail says in Issue #17:
my question is:
are the token between different dialogue turns the same as the tokens separating whole dialogue sessions?
Thank you
The text was updated successfully, but these errors were encountered: