Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update OpenXLA-pin to Nov24 #3

Closed
wants to merge 635 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
635 commits
Select commit Hold shift + click to select a range
a129c9b
[xla_compile][NFC] Extract the compilation and file-writing to a libr…
pizzud Nov 14, 2023
3e2b71a
[xla:gpu] NFC: Remove nested kernel namespace from custom kernels/fus…
ezhulenev Nov 14, 2023
5159949
Enable macOS Arm64 nightly builds
nitins17 Nov 14, 2023
548a766
Enable NVCC only for XLA's build_gpu_nvcc Kokoro job
tensorflower-gardener Nov 14, 2023
1e05b09
Reverts fbe8a546c55cfb4c43c763473eccba5a8bfc239c
hawkinsp Nov 14, 2023
d6fa3e4
[XLA] Serialize replica_count and num_partitions module properties to IR
cheshire Nov 14, 2023
b4479af
[XLA:GPU] Loosen up expectations on Int8 gemm test.
tensorflower-gardener Nov 14, 2023
f1f77cd
Error out if ptxas version < 11.8
akuegel Nov 14, 2023
c5800a1
Fix test gpu_aot_compilation_test
tensorflower-gardener Nov 14, 2023
e25203a
[XLA] [NFC] Unify graph rendering for fusion visualization and HTML v…
cheshire Nov 14, 2023
297fd57
PR #6964: [ROCM] fixing build brakes 23-11-13
pemeliya Nov 14, 2023
878d317
[XLA] Minor tweaks to memory-bound loop optimizer.
berkinilbeyi Nov 14, 2023
2a10123
Enable running passes for H100
tensorflower-gardener Nov 14, 2023
6843ab3
Integrate LLVM at llvm/llvm-project@ed86e740effa
tensorflower-gardener Nov 14, 2023
01cbfb3
Add nvml headers for Windows, based on https://github.com/openxla/xla…
ddunl Nov 14, 2023
abe6c2b
Use sharding propagation to infer dot/conv sharding given the operand…
tensorflower-gardener Nov 14, 2023
ccd828e
[xla:gpu] Disable gpu_aot_compilation_test in jitrt_executable_tests
anlunx Nov 14, 2023
3d46b55
Update TFRT dependency to use revision
tensorflower-gardener Nov 14, 2023
1be779f
[xla:gpu] NFC: Fix typo in filecheck based test
ezhulenev Nov 14, 2023
3a8a598
[stream_executor] NFC: Rename KernelArgsArrayBase to KernelArgs
ezhulenev Nov 14, 2023
e963478
Create script for generating `compile_commands.json`
ddunl Nov 14, 2023
e176b35
Prevent OOB indexing in StableHLO/MHLO ops.
GleasonK Nov 14, 2023
825a010
[xla:gpu] Add gpu-schedule-postprocessing pass.
bixia1 Nov 14, 2023
da8f498
Populate manual sharding for asynchronous instructions.
ZixuanJiang Nov 14, 2023
508c700
[XLA] Exclude async ops from elapsed times in MSA since we expect asy…
berkinilbeyi Nov 15, 2023
b3d902c
Fix a bug in sliced prefetching in which it allows slice start times …
sparc1998 Nov 15, 2023
182544f
[TSL] Change coordination service const std::string& arguments for ke…
hawkinsp Nov 15, 2023
dece2a6
[xla:gpu] Add name to custom kernels to improve logging and debugging
ezhulenev Nov 15, 2023
0528619
Integrate LLVM at llvm/llvm-project@5d6304f01742
tensorflower-gardener Nov 15, 2023
8fb606f
Update TFRT dependency to use revision
tensorflower-gardener Nov 15, 2023
544bb24
[XLA:GPU] Add helpers to calculate fused and unfused time separately.…
olegshyshkov Nov 15, 2023
8b53b4b
[XLA] Allow XLA to propagate the layouts into parameters without layo…
blakehechtman Nov 15, 2023
cbe8ee9
Enhance the sorting lambda to be strict weak ordering by using instru…
tensorflower-gardener Nov 15, 2023
b0172f9
Multi-threaded fusion priority evaluation.
jreiffers Nov 15, 2023
92542a5
Add a private API to allow setting layouts on jitted computations.
yashk2810 Nov 15, 2023
68841a1
[XLA] Fix typo in cost model in dot_handler.
Nov 15, 2023
e64f453
Integrate LLVM at llvm/llvm-project@b88308b1b481
tensorflower-gardener Nov 15, 2023
5cd1bc0
Remove an unused variable.
changhuilin Nov 15, 2023
c7aa3bd
[xla:ffi] Add API for accessing platform specific stream in XLA FFI h…
ezhulenev Nov 15, 2023
550561b
[XLA] hlo-opt tool: conversions from HLO to optimized HLO/LLVM/PTX/TT…
cheshire Nov 15, 2023
00e2c36
Update TFRT dependency to use revision
tensorflower-gardener Nov 15, 2023
6f3bedf
Integrate LLVM at llvm/llvm-project@c5dd1bbcc37e
tensorflower-gardener Nov 15, 2023
f937353
Add --expt-extended-lambda and --expt-relaxed-constexpr options for N…
tensorflower-gardener Nov 15, 2023
d9e9b3b
Allow CompilationEnvironments::ProcessNewEnvFn to return error status.
tensorflower-gardener Nov 15, 2023
72322de
Turn on clang+nvcc compiler for rbe_linux_cuda_nvcc config.
tensorflower-gardener Nov 15, 2023
9f276d9
Update TFRT dependency to use revision
tensorflower-gardener Nov 15, 2023
e5a9465
When respecting user sharding annotations, consider all sharding stra…
tensorflower-gardener Nov 15, 2023
5f55b02
[HloValueSemanticsAnalysis] Adds handler for RNG bit generator.
tensorflower-gardener Nov 15, 2023
94770aa
[xla:gpu] Add optional custom fusions pass to XLA:GPU pipeline
ezhulenev Nov 15, 2023
5165ddd
Reverts 1e05b098213cca516b52d845f55dc9c7d34c0bb0
hawkinsp Nov 15, 2023
c65de64
Add `--output=jsonproto` to example usage in generate_compile_command…
ddunl Nov 15, 2023
d236e87
[xla:gpu] Add support for multi-instruction custom fusions
ezhulenev Nov 15, 2023
1c42852
Wrap GPU client creation options in a struct, and overload GetStreamE…
changhuilin Nov 15, 2023
d2db247
Turn computing default solutions using sharding propagation off by de…
tensorflower-gardener Nov 15, 2023
d698e8c
[xla:gpu] Add a stub for gemm with upcast custom fusion
ezhulenev Nov 15, 2023
18b9994
[stream_executor] Fix absl::Span usage in GpuCommandBuffer #5857
anlunx Nov 15, 2023
880b39b
Reintroduces some CP-SAT solver parameters that are designed to ensur…
tensorflower-gardener Nov 15, 2023
88af93f
[XLA] Match other dimensions to try to swap instead of just directly …
Nov 16, 2023
ae7703a
[PJRT] Fix PjRtExecutable::GetParameterLayouts for tupled params
skye Nov 16, 2023
dd7b496
[XLA:GPU] Modify the the triton softmax emitter to accept multiple pa…
jvstokes Nov 16, 2023
05ff81a
Replaces 'StrategyVector' with 'StrategyGroup'.
tensorflower-gardener Nov 16, 2023
a649b1b
Renames the list of strategy choices to simply 'strategies' (instead …
tensorflower-gardener Nov 16, 2023
8ad65a6
Remove a couple dead functions
tensorflower-gardener Nov 16, 2023
b50f7be
Integrate LLVM at llvm/llvm-project@8ea8dd9a0171
tensorflower-gardener Nov 16, 2023
ded2b9e
Update TFRT dependency to use revision
tensorflower-gardener Nov 16, 2023
f3cd8e1
Remove usage of deprecated dyn_cast, cast and isa member functions of…
akuegel Nov 16, 2023
d39da54
[XLA:GPU][NFC] Add a TritonTest base class in tests.
bchetioui Nov 16, 2023
c416ef6
Priority fusion: cache HloFusionAnalyses.
jreiffers Nov 16, 2023
bf6dd45
Refactor FloatSupport to allow configuring the high precision type.
akuegel Nov 16, 2023
0602659
Reverts c416ef6d4f65458bfa9fa8a4c6fb67fdacfe7fb5
jreiffers Nov 16, 2023
9f357a5
[XLA:GPU] Fix fusion parameter limit
tdanyluk Nov 16, 2023
ea87d1a
[XLA:GPU] Add GetCommonUtilization helper. (NFC)
olegshyshkov Nov 16, 2023
d467c70
Add Bazel toolchain configs for cross-compiling TensorFlow for Linux …
nitins17 Nov 16, 2023
8af9c3c
Integrate LLVM at llvm/llvm-project@865f54e50173
tensorflower-gardener Nov 16, 2023
0086931
Reverts 4a06c211ba4edc3706dee01073e63426cb2ca859
ddunl Nov 16, 2023
e5512b6
Update TFRT dependency to use revision
tensorflower-gardener Nov 16, 2023
71f2a3c
Clean up generate_compile_commands.py
ddunl Nov 16, 2023
c9fbf07
[PJRT C API] Fix pybind11 implicit conversion from boolean to integra…
sagyakwa Nov 16, 2023
4c3058e
Renames 'LeafStrategies' to 'StrategyGroups' (since each of its membe…
tensorflower-gardener Nov 16, 2023
48b9c3c
py::cast<absl::Span<int64_t>> forms a reference to a temporary inside…
pschuh Nov 16, 2023
80b5bd4
Symlink `bazel-xla/external` to `external` in generate_compile_comman…
ddunl Nov 16, 2023
eb3510d
Roll forward, skip shape inference for Gather with unbounded dynamism…
ghpvnist Nov 16, 2023
45980c2
Fix symlink in generate_compile_commands.py and only create symlink w…
ddunl Nov 16, 2023
d3af4cd
Reports the model fingerprint & solution fingerprint to help diagnose…
tensorflower-gardener Nov 16, 2023
1de6a00
Sorts the elements of the liveness (node) set to ensure the consisten…
tensorflower-gardener Nov 16, 2023
1d0ecec
[PJRT] Add utilities for attaching layout mode info to HLO module.
skye Nov 16, 2023
1fd255a
Rollback https://github.com/openxla/xla/commit/15177468f898efe4b27a1b…
reedwm Nov 17, 2023
814a1d3
Re-enable layering_check for target.
tensorflower-gardener Nov 17, 2023
df06500
Integrate LLVM at llvm/llvm-project@46396108deb2
tensorflower-gardener Nov 17, 2023
cc95116
[xla:gpu] Use xla_gpu_enable_command_buffer flag to control command b…
anlunx Nov 17, 2023
3d7c681
Catch all python errors in custom call sharding and convert to FATAL …
pschuh Nov 17, 2023
f7b7ce1
Update TFRT dependency to use revision
tensorflower-gardener Nov 17, 2023
9b3453a
Reverts cc95116423fde23a4fe611ceb2e889dd1317119c
liangyaning33 Nov 17, 2023
c083faa
Integrate LLVM at llvm/llvm-project@00da5eb86ed0
tensorflower-gardener Nov 17, 2023
d83b25d
Update TFRT dependency to use revision
tensorflower-gardener Nov 17, 2023
275060c
PR #7025: [NVIDIA XLA] Add 2 new attributes to hlo instructions to sp…
Tixxx Nov 17, 2023
4e71659
[xla:python] Don't assume GPU clients are PJRT-compatible.
chr1sj0nes Nov 17, 2023
5fe2240
Remove is_scheduled from .hlo files.
akuegel Nov 17, 2023
a70ff53
Priority fusion: cache HloFusionAnalyses.
jreiffers Nov 17, 2023
433831d
Integrate LLVM at llvm/llvm-project@de176d8c5496
tensorflower-gardener Nov 17, 2023
6afdb0f
Import openai/triton from GitHub.
tensorflower-gardener Nov 17, 2023
f1a7201
Update TFRT dependency to use revision
tensorflower-gardener Nov 17, 2023
5f3417f
Integrate LLVM at llvm/llvm-project@ec42d547eba5
tensorflower-gardener Nov 17, 2023
ae451ba
Update TFRT dependency to use revision
tensorflower-gardener Nov 17, 2023
592ebe3
[PJRT C API] Add non_donatable_input_indices to ExecuteOption.
Nov 17, 2023
610dd5b
[PJRT] NFC: Cleanups to PJRT CPU client.
hawkinsp Nov 17, 2023
b1ab18b
Represent int4 as the LLVM type i4 instead of i8.
reedwm Nov 17, 2023
20d7a62
Fix flaky scatter3d test. In the test we only constrain the memory bu…
tensorflower-gardener Nov 17, 2023
f6252ec
Replace dependency on xla/util.h with dependency on math_util.h.
tensorflower-gardener Nov 17, 2023
6dbbc1c
[XLA:GPU] Expose buffer-assignment from hlo-opt
cheshire Nov 17, 2023
1e1321b
PR #7054: Incresed the tolerance for Ampere
hmonishN Nov 17, 2023
cb6b3a4
[xla:gpu] Rename buffer assignment getters to fix compilation error
ezhulenev Nov 17, 2023
af885c2
Move custom device trace up by reorder the pid allocation to be (a to…
zzzaries Nov 17, 2023
42b7727
Fix crash with int4 scalar arrays.
reedwm Nov 17, 2023
871e2a6
[stream_executor] Add efficient command buffer updates
tyb0807 Nov 17, 2023
875b014
[xla:python] Defer evaluating `platform_name` in `PyClient`.
chr1sj0nes Nov 17, 2023
35b361c
Fix some compilation perf issues in copy_insertion.cc
tensorflower-gardener Nov 17, 2023
31f68db
Reorganize `hlo_test` to only take one input HLO
ddunl Nov 17, 2023
077db88
[xla:gpu] Revise gpu-schedule-postprocessing pass to exclude fusion.
bixia1 Nov 18, 2023
ba6a538
Add UnTileShape() and UnTileLeafShape() in hlo_sharding_util.
tensorflower-gardener Nov 18, 2023
f787526
[XLA] Fix algebraic simplifier to not mess with async DUS/DS.
Nov 18, 2023
f1f2835
Reverts 9b3453a6812777d4d9d31296db443fa37fb1090a
anlunx Nov 18, 2023
335881b
[xla:gpu] Add no_parallel_custom_call to LMHLO_GPU async collective o…
bixia1 Nov 18, 2023
8762b61
[xla:ffi] Add support for int64_t attributes to XLA FFI
superbobry Nov 18, 2023
59e496a
[xla:ffi] Add API for accessing a variable number of arguments in an …
superbobry Nov 18, 2023
85c2d8c
Lazily compute `xla::Executable::hlo_proto()` to reduce memory consum…
tensorflower-gardener Nov 18, 2023
9df8e4b
[XLA] Move `CopyDenseElementsDataToXlaFormat` closer to the ir_emitte…
tensorflower-gardener Nov 18, 2023
c4a813e
[xla:ffi] Add basic public FFI API and a test
ezhulenev Nov 18, 2023
a01af1a
Make the SpecifiedLayout class opaque.
yashk2810 Nov 18, 2023
a778d08
[xla:ffi] Add DataType enum to C API
ezhulenev Nov 19, 2023
85213eb
[XLA:GPU] [NFC] Better error messages when failing Triton autotuning
cheshire Nov 19, 2023
6416c78
PR #6905: [NVIDIA XLA GPU] Expose collective matmul trigger in debug …
Tixxx Nov 20, 2023
775f1be
Integrate LLVM at llvm/llvm-project@506c47df00bb
tensorflower-gardener Nov 20, 2023
3f905cb
PR #7088: Fix incorrect device memory limit on non root processes.
trevor-m Nov 20, 2023
c43d460
Update TFRT dependency to use revision
tensorflower-gardener Nov 20, 2023
9cf07ca
PR #7029: [ROCm] add fpus for rocm and enable xla_compile_lib_test
i-chaochen Nov 20, 2023
3ec343e
Remove is_scheduled from .hlo file.
akuegel Nov 20, 2023
989d262
PR #7132: adding fix for local_client_aot_test target build failure
hmonishN Nov 20, 2023
661b80b
[XLA:GPU] Fix type of pred parameters in Triton fusion
tdanyluk Nov 20, 2023
87b89f6
Deduplicate algorithms in CreateOpRunners()
akuegel Nov 20, 2023
80d5e9d
Reverts a70ff534d086665dc35558d94f590045068ddcf9
jreiffers Nov 20, 2023
46d6359
[XLA:GPU][NFC] Rename kParameterPerScope to kParameterPerDotScope.
bchetioui Nov 20, 2023
6d05037
[XLA:GPU] Use the same coalescing heuristics for fused and unfused case.
olegshyshkov Nov 20, 2023
274fbb7
[XLA:GPU] Remove unnecessary local variable.
dimitar-asenov Nov 20, 2023
c4c27ce
[XLA:GPU] Add GetOperandUtilization helper. (NFC)
olegshyshkov Nov 20, 2023
dc3f78a
[xla:gpu] Extend LMHLO to GPU runtime lowering to the no_parallel_cus…
bixia1 Nov 20, 2023
9923428
Disable MMA_V3 in triton by default
tensorflower-gardener Nov 20, 2023
237db65
Integrate StableHLO at openxla/stablehlo@444c72d6
GleasonK Nov 20, 2023
04e123e
[XLA] [NFC] Expose RenderGraph in dump.h header for outside usage
cheshire Nov 20, 2023
4b00fdb
HloFusionAnalysis: Don't store references to ops outside the analyzed…
jreiffers Nov 20, 2023
7166d4b
Lower tensor.dim to MHLO in ShapeLegalizeToHLO
tensorflower-gardener Nov 20, 2023
fe048a7
[XLA:GPU] Use LazyRE2 so regexes don't show up in compile time profiles
d0k Nov 20, 2023
de6af0e
Prevent all `reduce-reduce` fusions.
jreiffers Nov 20, 2023
d8fc279
[XLA] Make complex resharding more aggressive.
Nov 20, 2023
76bc400
Integrate LLVM at llvm/llvm-project@9bdbb8226e70
krasimirgg Nov 20, 2023
c037ba2
[XLA] Avoid making an extra copy of constant literals when compiling …
dimitar-asenov Nov 20, 2023
26f63f5
Update TFRT dependency to use revision
tensorflower-gardener Nov 20, 2023
1f997b7
Add a default constructor for ffi::Span
superbobry Nov 20, 2023
cb9ceea
[xla:gpu:runtime] Support `GetObjFile` for non-`JitExecutable`.
chr1sj0nes Nov 20, 2023
11a4d8d
[Triton] Fix a dominance violation bug in `TritonGPURemoveLayoutConve…
tensorflower-gardener Nov 20, 2023
4273684
[stream_executor] Add support for conditional If command #6973
ezhulenev Nov 20, 2023
e68be78
[xla:ffi] Add support for passing dictionaries in attributes
ezhulenev Nov 20, 2023
053134a
Internal Code Change
ckennelly Nov 20, 2023
f1886f5
[xla:gpu] Enable command buffers if thunk runtime is enabled #6528
anlunx Nov 21, 2023
ef241af
Skip failing test temporarily in macOS Arm64 continuous builds
nitins17 Nov 21, 2023
85f07af
[xla:gpu] Extend p2p-schedule-preparation pass to also order custom-c…
bixia1 Nov 21, 2023
e142d2c
[xla:gpu] Use no_parallel_custom_call attribute to enable clique opti…
bixia1 Nov 21, 2023
c427d3d
[XLA] Only relayout literals if necessary during LayoutAssignment.
dimitar-asenov Nov 21, 2023
d01aff3
[TileAnalysis] Add indexing computations for several most common ops.
pifon2a Nov 21, 2023
1f3d0a8
XRT is no longer in use. Remove the code.
blakehechtman Nov 21, 2023
9dd3175
[XLA:GPU] Create the typedef OldToNewHloMap
tdanyluk Nov 21, 2023
58e1578
[XLA:GPU] [NFC] Add HTML generation capabilities to hlo-opt
cheshire Nov 21, 2023
5b4305a
Integrate LLVM at llvm/llvm-project@ec1086f2a085
krasimirgg Nov 21, 2023
9f103a7
[xla:gpu] Put CUDA graph conditional under an explicit define
ezhulenev Nov 21, 2023
e75dd08
Update TFRT dependency to use revision
tensorflower-gardener Nov 21, 2023
e4eabf7
[xla:ffi] Add support for decoding attributes into structs
ezhulenev Nov 21, 2023
f4eda48
[XLA:GPU] Allow overriding the external thread pool for autotuning too
tdanyluk Nov 21, 2023
ca82fcc
Integrate StableHLO at openxla/stablehlo@2e78612
sdasgup3 Nov 21, 2023
a948d5f
Update to use the xla::GetStreamExecutorGpuClient() that takes a stru…
changhuilin Nov 21, 2023
c32e76f
[xla:ffi] Add performance benchmarks for XLA:FFI
ezhulenev Nov 21, 2023
70c2e4c
Parameterized ffi::BaseBuffer with its dtype
superbobry Nov 21, 2023
a0df42b
Add enable_stub_generation parameter pybind_extension macros.
viswanadha9 Nov 21, 2023
b08e146
Split ParseFlagsFromEnvAndDieIfUnknown into 2 independent functions.
tensorflower-gardener Nov 21, 2023
79034f9
Integrate LLVM at llvm/llvm-project@56f569049c9d
tensorflower-gardener Nov 22, 2023
162dfa9
[xla:gpu] Create HLO version of EmitLmhloRegion and EmitOp, and use t…
anlunx Nov 22, 2023
97e4c98
Update TFRT dependency to use revision
tensorflower-gardener Nov 22, 2023
2517bd4
Fix case for Infeed handling in HloVerifier.
tensorflower-gardener Nov 22, 2023
593d04b
[xla:gpu] Update conditional command buffer for If branch #6973
ezhulenev Nov 22, 2023
d92962b
[xla:gpu] NFC: remove EntryFunctionAttributes
anlunx Nov 22, 2023
2671a3f
Apply ClangTidy fixes to change down_cast to proto2::DownCastToGenera…
tensorflower-gardener Nov 22, 2023
ae24a41
PR #7118: [ROCm] enabled gpu_aot_compilation_test for ROCm
weihanmines Nov 22, 2023
6751712
PR #7179: [XLA:GPU] add back pointer in while body computation to tra…
Cjkkkk Nov 22, 2023
78f3c5a
PR #7134: [XLA:GPU] Instrument LowerToXlaGpuRuntime to dump IR after …
andportnoy Nov 22, 2023
14e518b
Replace the HLO and MHLO TopK `comparator` with a simpler boolean arg…
dimitar-asenov Nov 22, 2023
fa7c1db
[XLA] Add `XlaBuilder::TopK`
dimitar-asenov Nov 22, 2023
e8bdc39
PR #7189: [ROCm] Fixed rocm build for conditional graph
i-chaochen Nov 22, 2023
2172ac6
[XLA:GPU] Fuse the scopes of the dot fusion to separate subgraphs
tdanyluk Nov 22, 2023
e521aa2
[XLA] Add `mhlo.topk` -> `hlo.topk` translation.
dimitar-asenov Nov 22, 2023
0eace63
PR #7197: [ROCm] add navi31 on xla
i-chaochen Nov 22, 2023
863d1d1
Integrate LLVM at llvm/llvm-project@cc2128714e66
krasimirgg Nov 22, 2023
67cda44
Update TFRT dependency to use revision
tensorflower-gardener Nov 22, 2023
95a8715
[XLA:GPU] priority_fusion: prevent fusions that break epilogue fusions
d0k Nov 22, 2023
393afac
[XLA:GPU] Remove now unnecessary filter gating Softmax rewrites.
bchetioui Nov 22, 2023
cc5307e
Do not allow reduction epilogue fusion for reductions that require at…
akuegel Nov 22, 2023
dd5419f
Reverts 80d5e9d6c2a25f0aa0c93f7f81e366cf59516a05
jreiffers Nov 22, 2023
2df495a
Add CUDA 12.3 symbols to TSL symbol files
tensorflower-gardener Nov 22, 2023
5140fc8
[XLA:GPU] Clean up the buffer assignment passing when creating a `Gpu…
dimitar-asenov Nov 22, 2023
36a2a06
[XLA:GPU] Add proper calculation of memory usage and compute time.
olegshyshkov Nov 22, 2023
c010566
Integrate LLVM at llvm/llvm-project@2b0b0ad7e760
krasimirgg Nov 22, 2023
08701b2
[XLA] hlo-opt: fix default stage value.
olegshyshkov Nov 22, 2023
b0ec7ba
Add DCN collective stat information as stat metadata to XPlane
SurbhiJainUSC Nov 22, 2023
cc6d188
Disable `//tests:lax_test_gpu` in JAX build script because it is curr…
ddunl Nov 22, 2023
0acb46f
[xla:gpu] Track command buffer state explicitly for each executor
ezhulenev Nov 22, 2023
8081622
[xla:gpu] Do not keep StreamExecutor pointer in CommandBuffer
ezhulenev Nov 22, 2023
656d5e5
Integrate StableHLO at openxla/stablehlo@e3276cd
sdasgup3 Nov 22, 2023
b2bb561
Use appropriate RBE config in JAX GPU build script
ddunl Nov 22, 2023
235c05c
Reenable `//tests:lax_test_gpu` in JAX build script, disable specific…
ddunl Nov 22, 2023
50f4d49
[xla:gpu] Add AotCompilationResult for thunk runtime
anlunx Nov 23, 2023
6be4202
Remove the deprecated PJRT GPU client creation.
changhuilin Nov 23, 2023
1235734
PR #7222: [ROCM] fixes for hipblas-lt 6.0 enums and datatypes
pemeliya Nov 23, 2023
9c92c6a
[xla:gpu] Add support for conditional commands
ezhulenev Nov 23, 2023
d17c9dd
[stream_executor] NFC: Use build flags to enable/compile CUDA graph c…
ezhulenev Nov 23, 2023
b6026b8
PR #7034: [ROCM] enabling gemm_algorithm_picker on ROCM platform
pemeliya Nov 23, 2023
d59eb4d
PR #7016: [NVIDIA XLA GPU] Restore dotdimension numbers for original …
Tixxx Nov 23, 2023
aa85a3d
[XLA] Add `hlo.topk` -> `mhlo.topk` translation.
dimitar-asenov Nov 23, 2023
ae906af
Update XLA's Triton pipeline with the missing changes from the latest…
gflegar Nov 23, 2023
3803c88
[xla:gpu] Allow missing ObjFile / MLIR in `GpuCompiler::Export`.
chr1sj0nes Nov 23, 2023
0cfb44a
[XLA] Disable autosharding tests in OSS until it's fully fixed upstream
cheshire Nov 23, 2023
b37ad4b
PR #7201: [GPU] Improve NCCL error reporting
nouiz Nov 23, 2023
6cb8452
PR #7237: Update build_from_source.md doc to point to latest-gpu dock…
joker-eph Nov 23, 2023
a1dd132
Fix a typo in comment.
penpornk Nov 23, 2023
30de7d8
[TileAnalysis] Use IsOpElementwise instead of IsElementwise.
pifon2a Nov 23, 2023
a53fcaa
Add MoveCopyToUsers pass to multi-headed attention fusion pipeline.
akuegel Nov 23, 2023
7239eb4
[XLA:GPU] Tiled fusion: fix nested slicing.
Nov 23, 2023
2f88516
Import openai/triton from GitHub.
tensorflower-gardener Nov 23, 2023
69a3c3b
[XLA:GPU] Do not error out in triton autotuner in deviceless mode whe…
cheshire Nov 23, 2023
a191d92
[XLA:GPU] Fix occasional nullptr dereference in IrEmitterUnnested::Em…
tdanyluk Nov 23, 2023
1dba8f6
[xla:ffi] Sketched basic diagnostics infrastructure and plumbed it th…
superbobry Nov 23, 2023
f6856c2
[stream_executor] Add If-Else conditional commands to CommandBuffer
ezhulenev Nov 23, 2023
1fe1c45
Allow unbounded dynamism representation for AddOp.
ghpvnist Nov 24, 2023
af6dbb7
Adding fingerprint to module's profile information
tensorflower-gardener Nov 24, 2023
7e2c601
Integrate LLVM at llvm/llvm-project@af7a1453526a
krasimirgg Nov 24, 2023
a3bfe5d
Update TFRT dependency to use revision
tensorflower-gardener Nov 24, 2023
eb15c02
[TileAnalysis] Add indexing computation for fusionOp.
pifon2a Nov 24, 2023
fbe3de1
Replace boundary functions with a fusion adaptor class.
jreiffers Nov 24, 2023
0ad1d72
[XLA:GPU][TileAnalysis] Add support for dot in tile analysis.
bchetioui Nov 24, 2023
e5a0ad2
Move vectorize_copy.cc and copy_removal.cc passes out of gml_st.
pifon2a Nov 24, 2023
acd327d
[TileAnalysis] Add a test for dynamic-update-slice.
pifon2a Nov 24, 2023
8744c9a
Put thlo & gml_st on ice.
pifon2a Nov 24, 2023
e826787
Merge commit '8744c9a94782cd7804f015e6d29df253437af3cb' into HEAD
Dec 5, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
151 changes: 113 additions & 38 deletions .bazelrc

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion .github/workflows/trusted_partners.js
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,7 @@ const get_email_domain = async ({github, username}) => {
const filter_action = async ({github, context, domain}) => {
const labels = ['kokoro:force-run'];

let assignees = ['radhakrishnaba', 'xla-rotation'];
let assignees = ['kamaljeeti', 'xla-rotation'];
const title =
context.payload.pull_request && context.payload.pull_request.title;
const lowercased_title = (title || '').toLowerCase();
Expand Down
108 changes: 108 additions & 0 deletions .kokoro/jax/build.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,108 @@
#!/bin/bash
# Copyright 2022 Google LLC All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

# -e: abort script if one command fails
# -u: error if undefined variable used
# -o pipefail: entire command fails if pipe fails. watch out for yes | ...
# -o history: record shell history
set -euox pipefail -o history

# Builds + tests jaxlib against CL/PR version of XLA + JAX main.

source "${KOKORO_GFILE_DIR}/utils.sh"

function is_linux_gpu_job() {
[[ "$KOKORO_JOB_NAME" =~ tensorflow/xla/jax/.*gpu.* ]]
}

clone_main_jax() {
git clone https://github.com/google/jax.git
}

prelude() {
export JAX_ENABLE_X64=0

if is_linux_gpu_job ; then
export JAX_CUDA_VERSION=12
export JAX_CUDNN_VERSION=8.9
nvidia-smi
setup_env_vars_py39
else
setup_env_vars_py312
fi

cd "${KOKORO_ARTIFACTS_DIR}"

use_local_or_install_python
install_packages "$NUMPY_VERSION" "$SCIPY_VERSION"
clone_main_jax
# Install bazel
update_bazel_linux

cd jax

}

build_and_test_on_rbe_cpu() {
# Run the tests.
bazel \
test \
--verbose_failures=true \
--override_repository=xla="${KOKORO_ARTIFACTS_DIR}"/github/xla \
--config=avx_posix \
--config=mkl_open_source_only \
--config="rbe_cpu_linux_py312" \
--config=tensorflow_testing_rbe_linux \
--test_env=JAX_NUM_GENERATED_CASES=25 \
--test_output=errors \
-- //tests:cpu_tests //tests:backend_independent_tests
}

build_and_test_on_rbe_gpu() {
# Runs non-multiaccelerator tests with one GPU apiece.
# It appears --run_under needs an absolute path.

# TODO(ddunleavy): reenable `LaxTest.testBitcastConvertType`
bazel \
test \
--verbose_failures=true \
--override_repository=xla="${KOKORO_ARTIFACTS_DIR}"/github/xla \
--config=avx_posix \
--config=mkl_open_source_only \
--config="rbe_linux_cuda12.2_nvcc_py3.9" \
--config=tensorflow_testing_rbe_linux \
--test_env=XLA_PYTHON_CLIENT_ALLOCATOR=platform \
--test_output=errors \
--test_env=JAX_SKIP_SLOW_TESTS=1 \
--test_env=TF_CPP_MIN_LOG_LEVEL=0 \
--test_env=JAX_EXCLUDE_TEST_TARGETS="PmapTest.testSizeOverflow|LaxTest.testBitcastConvertType" \
--test_tag_filters=-multiaccelerator \
-- //tests:gpu_tests //tests:backend_independent_tests
}

# Generate a templated results file to make output accessible to everyone
"$KOKORO_ARTIFACTS_DIR"/github/xla/.kokoro/generate_index_html.sh "$KOKORO_ARTIFACTS_DIR"/index.html

prelude

if is_linux_gpu_job ; then
build_and_test_on_rbe_gpu
else
build_and_test_on_rbe_cpu
fi

echo "bazel-testlogs (test results) location:"
find "$KOKORO_ARTIFACTS_DIR" \
-type l,d -name bazel-testlogs || echo "bazel-testlogs not found"
17 changes: 13 additions & 4 deletions .kokoro/linux/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,10 @@ function is_linux_gpu_job() {
[[ "$KOKORO_JOB_NAME" =~ tensorflow/xla/linux/.*gpu.* ]]
}

function is_use_nvcc() {
[[ "${USE_NVCC:-}" == "true" ]]
}

# Pull the container (in case it was updated since the instance started) and
# store its SHA in the Sponge log.
docker pull "$DOCKER_IMAGE"
Expand All @@ -44,16 +48,23 @@ RC_FILE="/usertools/cpu.bazelrc"
TARGET_FILTER=""
TAGS_FILTER="-no_oss,-oss_excluded,-oss_serial"
ADDITIONAL_FLAGS=""
RBE_CONFIG=""

if is_linux_gpu_job ; then
TAGS_FILTER="$TAGS_FILTER,gpu,requires-gpu-nvidia,-no_gpu"
ADDITIONAL_FLAGS="$ADDITIONAL_FLAGS --run_under=//tools/ci_build/gpu_build:parallel_gpu_execute"
RC_FILE="/usertools/gpu.bazelrc"
if is_use_nvcc ; then
RBE_CONFIG="rbe_linux_cuda_nvcc"
else
RBE_CONFIG="rbe_linux_cuda"
fi
echo "***NOTE: nvidia-smi lists the highest CUDA version the driver supports, which may be different than the version of CUDA actually used!!***"
nvidia-smi
else
TAGS_FILTER="$TAGS_FILTER,-gpu,-requires-gpu-nvidia"
ADDITIONAL_FLAGS="$ADDITIONAL_FLAGS --config=nonccl"
RBE_CONFIG="rbe_linux_cpu"
fi

# Build & test XLA
Expand All @@ -65,7 +76,7 @@ docker exec xla bazel --bazelrc=$RC_FILE \
--features=layering_check \
--profile=/tf/pkg/profile.json.gz \
--flaky_test_attempts=3 \
--config=rbe \
--config=$RBE_CONFIG \
--jobs=150 \
--nobuild_tests_only \
$ADDITIONAL_FLAGS \
Expand All @@ -74,9 +85,7 @@ docker exec xla bazel --bazelrc=$RC_FILE \

# Print build time statistics, including critical path.
docker exec xla bazel analyze-profile "/tf/pkg/profile.json.gz"
# TODO(ddunleavy): enable once container has clang-tidy
# docker exec xla git config --global --add safe.directory /tf/xla
# docker exec xla bash -c "bazel aquery --output=jsonproto \"mnemonic(CppCompile, //xla/...)\" | PYTHONPATH=.. python3 build_tools/lint/clang_tidy.py --changed_lines_only"

# Stop container
docker stop xla

Expand Down
17 changes: 6 additions & 11 deletions build_tools/lint/BUILD
Original file line number Diff line number Diff line change
Expand Up @@ -13,24 +13,14 @@
# limitations under the License.
# ============================================================================

load(
"//xla:pytype.default.bzl",
"pytype_strict_binary",
"pytype_strict_library",
)
load("//xla:pytype.default.bzl", "pytype_strict_binary", "pytype_strict_library")
# Placeholder: load py_test

package(
# copybara:uncomment default_applicable_licenses = ["//tensorflow:license"],
licenses = ["notice"],
)

pytype_strict_binary(
name = "clang_tidy",
srcs = ["clang_tidy.py"],
deps = [":diff_parser"],
)

pytype_strict_library(
name = "check_contents",
srcs = ["check_contents.py"],
Expand All @@ -43,6 +33,11 @@ pytype_strict_library(
visibility = ["//visibility:public"],
)

pytype_strict_binary(
name = "generate_compile_commands",
srcs = ["generate_compile_commands.py"],
)

py_test(
name = "check_contents_test",
srcs = ["check_contents_test.py"],
Expand Down
Loading
Loading