-
Notifications
You must be signed in to change notification settings - Fork 534
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LORA + FSDP: ValueError: FlatParameter
requires uniform requires_grad
#414
Comments
Hi @minionKP. We do not yet have LoRA working with FSDP. To use LoRA use DDP. We expect to add support eventually, but do not have a solid timeline. |
Want to do LORA on MPT-30B model. From another issue #413 looks like QLORA or doing |
I think this issues not only arises with LoRA but with any mixed setup of frozen and unfrozen parameters. Is there any solution in sight? I know one could write a custom auto-wrap policy when using lightning or pure PyTorch but that seems to be complicated especially because it creates un-uniformly sized FSDPUnits :-/ |
Environment
To reproduce
Steps to reproduce the behavior:
Expected behavior
FSDP + LORA fine-tuning
Config file
No other changes to example config file
The text was updated successfully, but these errors were encountered: