-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix Mochi Quality Issues #10033
Merged
+337
−159
Merged
Fix Mochi Quality Issues #10033
Changes from all commits
Commits
Show all changes
55 commits
Select commit
Hold shift + click to select a range
27f81bd
update
DN6 30dd9f6
update
DN6 10275fe
update
DN6 79380ca
update
DN6 21b0997
update
DN6 fcc59d0
update
DN6 1782d02
update
DN6 66a5f59
update
DN6 3ffa711
update
DN6 dded243
update
DN6 d99234f
update
DN6 8b9d5b6
update
DN6 2cfca5e
update
DN6 900fead
update
DN6 0b09231
update
DN6 883f5c8
update
DN6 59c9f5d
update
DN6 f3fefae
update
DN6 8a5d03b
update
DN6 b7464e5
update
DN6 fb4e175
update
DN6 61001c8
update
DN6 0fdef41
update
DN6 e6fe9f1
update
DN6 c17cef7
update
DN6 0e8f20d
update
DN6 6e2011a
update
DN6 9c5eb36
update
DN6 d759516
update
DN6 7854bde
update
DN6 2881f2f
update
DN6 7854061
update
DN6 b904325
Merge branch 'main' into mochi-quality
sayakpaul ba9c185
update
DN6 53dbc37
update
DN6 77f9d19
update
DN6 a298915
Merge branch 'mochi-quality' of https://github.com/huggingface/diffus…
DN6 dc96890
update
DN6 ae57913
update
DN6 7626a34
update
DN6 c39886a
update
DN6 bbc5892
update
DN6 3c70b54
update
DN6 11ce6b8
update
DN6 cc7b91d
update
DN6 09fe7ec
Merge branch 'main' into mochi-quality
DN6 ccabe5e
Merge branch 'main' into mochi-quality
a-r-r-o-w 4c800e3
Merge branch 'main' into mochi-quality
DN6 2a6b82d
update
DN6 1421691
Merge branch 'mochi-quality' of https://github.com/huggingface/diffus…
DN6 cbbc54b
update
DN6 952f6e9
Merge branch 'main' into mochi-quality
DN6 b75db11
update
DN6 50c5607
Update src/diffusers/models/transformers/transformer_mochi.py
DN6 d80f477
Merge branch 'main' into mochi-quality
DN6 File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
OK! But cc @a-r-r-o-w here. He has been following the mochi-fix PR and added the attention processors to model files
I guess we keep them here for now until we refactor and move them all together?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we should do the following going forward as design choice (just personal opinion so let's try to forumalate a plan for consistency):
attention_processor.py
is now > 5k linesAttention
class is required, let's keep it in the transformer file as well. These custom classes require some common methods that will probably not change between implementantions. For this, let's create aAttentionMixin
class - for changing/getting attention processors, fusing, etc.functools.cache
works here if we make the rope calculation forward dependant on just the num_frames, height, width. But a save hook would work as well. This - we can look into a bit laterWDYT?