Update non_xla attention to properly support paged_attention dynamo code path #8072
build_and_test.yml
on: pull_request
Build PyTorch/XLA
/
build
26m 41s
Build XLA CUDA plugin
/
build
4m 44s
Build upstream Docker image
/
build
Matrix: CPU tests / test
Matrix: GPU tests / test
Artifacts
Produced during runtime
Name | Size | |
---|---|---|
cpp-test-bin
Expired
|
644 MB |
|
cuda-plugin
Expired
|
95.3 MB |
|
torch-xla-wheels
Expired
|
206 MB |
|