Skip to content

Fix flash attention GQA bug to use the dynamic size of the key/value tensors - used for eval/inference #3109

Fix flash attention GQA bug to use the dynamic size of the key/value tensors - used for eval/inference

Fix flash attention GQA bug to use the dynamic size of the key/value tensors - used for eval/inference #3109

Triggered via pull request November 21, 2023 23:12
Status Success
Total duration 6m 55s
Artifacts

code-quality.yaml

on: pull_request
Matrix: code-quality
Fit to window
Zoom out
Zoom in