Skip to content

Bumping flash attention version to 2.6.3 and adding option for softcap in attention and lm_head logits. #7164

Bumping flash attention version to 2.6.3 and adding option for softcap in attention and lm_head logits.

Bumping flash attention version to 2.6.3 and adding option for softcap in attention and lm_head logits. #7164

Annotations

1 error

This job was cancelled