Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[
core
] Integrate Flash attention 2 in most used models #25598[
core
] Integrate Flash attention 2 in most used models #25598Changes from 10 commits
8bb77a1
2e18421
fe5795e
4bd15e2
49fe318
f5d440b
50491e8
7df78c0
0e30d13
ad8b905
3c31f10
2628bf3
56d0b49
20d1b37
a82f1ca
c72e8ff
66823f9
41f8f3d
a64a1a9
67e3fc2
8444ab6
8b1c2df
c3ebcd2
1c212d8
4618701
85ec946
2248f20
0881ced
a8a1b2d
2be3e03
b6d3e58
b47e85c
db8bd64
58848ab
3f73557
baae736
55f6140
10d5c1b
3fb221a
a931aeb
68a1204
36e0d6e
2beeb68
7b5da2c
b99a582
adaed45
7f06af6
2d36c6f
a663fa4
9d3693f
65ae59c
43185b5
c61157e
2f17792
65c3861
165a503
5abc702
5069e4a
11400d8
ace7939
fe9b16d
6174c06
acfc954
33a0f62
ee8ba20
e28fb0b
025727c
8f7e400
3259392
57a077b
e62b0b8
7419438
3ba5e98
585e463
ec0f8b9
3e5ea35
4bb1bc5
3ea4633
b67c21e
5b73557
48e3bcf
0461384
8d72a66
73b2f07
fb7654c
a737bde
80951ae
6f7ff42
257a633
360da70
1d91bc4
8ecab97
7c5720f
28b82e2
84b5793
949172f
825c7e0
1af232c
d7f16c5
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'd add a link to https://github.com/huggingface/transformers/issues/new
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe print the current version they have installed currently
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Makes sense!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
should use
module.named_children()
instead ofmodule.named_modules()
named_modules()
will return self module with blank nameThere was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is this is the place where we get the original weigts to the new module?