FlashAttention 2 implementing PagedAttention #7321
Unanswered
shashank2000
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi - it appears PagedAttention has been implemented in FlashAttention 2.5 recently. Is there a world in which integrating a new model with vllm is potentially easier resulting from this?
Beta Was this translation helpful? Give feedback.
All reactions