-
-
Notifications
You must be signed in to change notification settings - Fork 896
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mistral example fail ValueError: You are attempting to perform batched generation with padding_side='right' #677
Comments
Fixed in #681 . Please let me know if the issue still arises. |
I'm still seeing this issue after the merge. The mistral example trains without a problem after altering it use qlora, but when I switch datasets from the original alpaca_2k_test to a local sharegpt-formatted dataset, I get the error again. |
@NanoCode012 Yes +1 @wmiller , using the sharegpt format produces the same error as before, and the example (with lora for memory reasons) works. |
@wmiller @kvikk , if you're using local dataset, could you try clear your Secondly, is this limited to only sharegpt format + Mistral + FA combo? Just trying to figure out the common fault. Is it still the paddIng_side issue ? |
I cleaned the path, so the dataset was reprocessed. I only tried the sharegpt format (and the example with alpaca which works now). The error message is the same as above. |
To reproduce, this yml can be used: ( The dataset is just a random set from hf I found, but it is in sharegpt format. It throws a lot of "roles not alternating" warnings. Rename to yml) |
still happens, yes |
#728 |
Thank you! It does work now. |
Please check that this issue hasn't been reported before.
Expected Behavior
Training should work.
Current behaviour
Running the example fails with:
File "/home/axolotl/env/lib/python3.10/site-packages/transformers/models/mistral/modeling_mistral.py", line 882, in forward raise ValueError( ValueError: You are attempting to perform batched generation with padding_side='right' this may lead to unexpected behaviour for Flash Attention version of Mistral. Make sure to call
tokenizer.padding_side = 'left'before tokenizing the input.
Steps to reproduce
On the main branch (43856c0), install and run example
accelerate launch -m axolotl.cli.train examples/mistral/config.yml
req.txt
Config yaml
examples/mistral/config.yml
Possible solution
No response
Which Operating Systems are you using?
Python Version
3.10
axolotl branch-commit
main/43856c0a393fb7c4c44c56dc1a35ab7bc4bd52fd
Acknowledgements
The text was updated successfully, but these errors were encountered: