-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Sigmas] Keep sigmas on CPU #6173
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
149b92d
to
d20be08
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you!
* correct * Apply suggestions from code review * make style
* correct * Apply suggestions from code review * make style
I'm curious about how it works. |
to is inplace. |
Cc: @patrickvonplaten here. This still doesn't affect the results in the PyTorch blog post as it was done from a separate branch with another way of doing the device placement. |
ahh does this means our |
This should not be a problem, since |
* correct * Apply suggestions from code review * make style
What does this PR do?
After a second round of testing, it seems like keeping the sigmas always on CPU is actually not a bad idea because there is no real slow-down for the "default" mode while the speed-up for torch.compile is significantly higher when leaving sigmas on CPU.
cc @sayakpaul