You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey thanks for building this framework it is exactly what I need for my project but I was wondering, whether there is a particular reason for why flash attention 2 and rotary positional embeddings were discarded from the standard Llama implementation?
The text was updated successfully, but these errors were encountered:
hi @benjamin-kroeger, it might be because this code is based on an earlier transformers version. I will upgrade it to be compatible with the latest transformers when I am free. PRs are also welcome!
Hey thanks for building this framework it is exactly what I need for my project but I was wondering, whether there is a particular reason for why flash attention 2 and rotary positional embeddings were discarded from the standard Llama implementation?
The text was updated successfully, but these errors were encountered: