-
-
Notifications
You must be signed in to change notification settings - Fork 893
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
semi-weekly 8bit lora zero3 check #1852
base: main
Are you sure you want to change the base?
Conversation
ea0d058
to
4e64549
Compare
this also needs this upstream PR huggingface/transformers#32943 |
958d25c
to
97b4529
Compare
97b4529
to
613a217
Compare
this one needs a deeper dive into why the train loss is larger by an order of magnitude |
I also tried deepspeed 0.16.1, but that is blocked too for gradient accumulation issues, see #2154. Also trying deepspeed==0.16.1 with grad_accum=1, also results in train/loss of ~13 on L3-3B. |
upstream issue opened here bitsandbytes-foundation/bitsandbytes#1451 |
adds a check on our semi-weekly GHA against 8bit lora with ds zero3. Depends on huggingface/transformers#32943 to be merged.