Skip to content

Releases: feifeibear/long-context-attention

Version 0.4.2 is released on 19th Nov 2024

19 Nov 02:12
c004be7
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.4.1...0.4.2

0.4.1 is released on Nov 15th 2024

15 Nov 08:34
7bbaf56
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.4.0...0.4.1

0.4.0 version is released on Nov 15th 2024

15 Nov 08:16
1fb7f00
Compare
Choose a tag to compare

The 0.4.0 version starts to support FlashAttention V3 on Hopper GPUs. This version also has been tested on low-memory (24GB) GPUs.

What's Changed

New Contributors

Full Changelog: 0.3.5...0.4.0

0.3.5

19 Sep 15:22
5503194
Compare
Choose a tag to compare

What's Changed

Full Changelog: 0.3.3...0.3.5

0.3.2 released

19 Sep 14:57
a56c301
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.3.1...0.3.2

v0.3.1 released at 2024.09.14

14 Sep 03:32
5ac27d8
Compare
Choose a tag to compare

stripe_extract_local, basic_extract_local, zigzag_extract_local works for tensors dimension >=2.

v0.3 released on 27th August 2024!

27 Aug 08:13
4bfd585
Compare
Choose a tag to compare

upgrade flash_attn >= 2.6.0

v0.2 released on 24th June 2024!

25 Jun 03:07
13f5af8
Compare
Choose a tag to compare
  1. Ulysses supports T4 and V100.
  2. Updates some directory structures.

v0.1

11 Apr 08:32
80c8b46
Compare
Choose a tag to compare

Sequence parallel attention adopting a hybrid ulysses and ring attention approach.
Support GQA
Support QKV packed.