Skip to content

v0.55.0-rc2

Pre-release
Pre-release
Compare
Choose a tag to compare
@github-actions github-actions released this 16 Jan 02:07
· 5 commits to main since this release
ca6730b

Note

If you are installing from a release, please refer to the README, INSTALLATION instructions, and any other documentation packaged with the release, not on the main branch. There may be differences between the latest main and the previous release.

The changelog will now follow, showing the changes from last release.

This release was generated by the CI workflow https://github.com/tenstorrent/tt-metal/actions/runs/12800305479

📦 Uncategorized

  • Add noc read/write burst command support to CCL command kernel. Also add automated command lowering to these noc commands
  • MeshWorkload: Initial Implementation
  • [CCL] Fix padding issues
  • #15868: use a buffer's size when creating its CB in groupnorm
  • Fix trace region size
  • #0: Bump E2E perf threshold for host bound WH Resnet variants
  • Extract Device interface
  • Extend graph capture to include device information
  • Quick fix replacing Device* with IDevice in graph tracker
  • #0: Add unit_tests_ttnn_tensor to post-commit
  • Xuncai/ccl global sem
  • #16153: Add fused activations to input tensors
  • Remove ARCH_NAME specific includes from erisc_datamover_builder
  • remove unused function
  • [TT-Train] Updates related to the fixed matmul
  • [Llama3] Add max prefill chunk sizes for different model/device combinations
  • Add sharded sweeps identiy, neg, selu, abs
  • Handle padded shards in ttnn.convert_to_chw
  • #16492: Add new APIs for setting which sub_device_ids to stall on
  • #0: Track local_cb_size to ensure that remote cb config is correctly sent by FD
  • support keepdim for prod
  • #16225: Int32 support for abs
  • Sharded sweeps: prelu, softmax, sinh, softplus, relu_max and relu_min
  • Changing output channel size in the readme example
  • Fix double move in TTNN invoke_composite launch_op
  • Quick fix how to storage/access for devices in the DevicePool
  • Add native N-dimensional tiled-interleaved permute support when the tiles are now broken.
  • fix multi-iter in reduce scatter and adopt runtime arg overrider infra
  • [tt-train] Add linear regression ddp example
  • Remove eth_l1_address_params.h from device.cpp
  • Sharded sweeps: exp, exp2, expm1, erfc, erfinv, round, log
  • Fix ttnn.concat golden function when groups > 1
  • #16171: Assert that NCRISC NOC is idle at kernel end.
  • Remove eth_l1_address_params.h from tt_cluster.cpp and watcher
  • Remove dev_mem_map.h usage from watcher_device_reader.cpp
  • #14616: Remove ARCH_* ifdefs from tt_cluster.cpp
  • Add support for DRAM Prefetcher op
  • Resolve reduce-scatter-async sharded tensor correctness bug & hang
  • disable flaky t3k test
  • Remove "noc_parameters.h" from device.cpp
  • Remove restriction of input_nsticks_per_core % w == 0
  • Add tt-forge sweep for conv2d.
  • Remove noc header file inclusion from watcher_device_reader.cpp
  • Fix ttnn.from_torch for 0D/1D tensors with tile layout
  • Short list failing conv2d for forge sweeps
  • Remove halo from shard spec
  • Address issues of var & std
  • #16492: Remove sub_device_ids apis from various read/write functions throughout the stack
  • #6344: Update RoBERTa QA demo
  • Remove noc_parameters.h inclusion from ttnn
  • Resubmit #16339: parameterize dispatch_constants
  • #11512: Refactor bitwise sweeps, add bitwise sharded sweeps, modify t…
  • Update CODEOWNERS
  • Enable multi-core and fixing bfloat8 for untilize with unpadding
  • Set up targeting idle eth cores on BH - won't enable because of hang debug
  • Reorganize Print Pages Infrastructure
  • lower fabric erisc datamover eth context switching frequency when workload is running
  • Composite binary sweeps: gcd and lcm
  • Remove ARCH_NAME from host library code
  • [tt-train] Add nanogpt ddp mode
  • #16312: Fix full op to query physical shape for buffer volume
  • #16366: Changed default kernal_config_val for 32bit matmul
  • #16621: Add barriers at end of cq_dispatch_slave.cpp
  • Build wheels in models unit tests workflow
  • Mo/10234 eth dispatch profiling
  • Support subcoregrids in concat_heads
  • Build wheels in ttnn unit tests workflow because the tests need it and we forgot to put it in
  • #16590: profiler trace detection fix
  • #16503: Optimize CoreRangeSets for CBs and semaphores
  • Revert "#16621: Add barriers at end of cq_dispatch_slave.cpp"
  • Fix nightly stable diffusion tests
  • #0: Used github team for conv files
  • Sweeps: fixed abs, added acos and acosh sharded and non sharded
  • fix reduce scatter multi-link support bug
  • support i/p tensors of all dimensions/rank for prod operation
  • Create Infrastructure to exactly calculate L1 Memory Usage for Conv2D #15088
  • #12253: Implement Batch norm operation for inference mode
  • Port all experimental ops to compute_output_specs
  • #16443: Add a programming example of vecadd_multi_core and gtest
  • Enable to/from torch tests for 0D/1D tensors
  • Port all data movements ops to compute_output_specs
  • #15246: Add sweep tests for addcdiv, addcmul, rdiv, rsub, ceil
  • Fix build break
  • Logical sharding for input tensor and halo output
  • #16495: reduce grid for falcon7b mlp matmul
  • Stress NOC mcast test
  • [skip ci] Update subdevice doc
  • Read from and write to partial buffer regions for interleaved buffers where offset and size of specified buffer region are divisible by buffer page size
  • Fix resnet large on GS
  • Fix Pre-allgather Layernorm bad PCC when use 1D reduction
  • #16353: skip no volume tensors
  • Create README.md
  • Update README.md
  • #16367: Added support to enable dram and l1 memory collection without saving to disk
  • Update .clang-format-ignore
  • Tweak BH csrrs init code
  • #0: Clean up confusing refs to Greyskull from ttnn.copy error messages.
  • Update perf and latest features for llm models (Jan 13)
  • Update README.md
  • #16657: Fix to_layout conversion into row major for 1D tensors
  • Tilize with val padding results in L1 cache OOM
  • #0: Fixes from commit ae61802
  • #0: Skip build-docker-image during post-commit code-analysis since the docker image is already built in a previous job
  • Generate test executables per architecture
  • #16587: Update UMD submodule commit for P150 compatibility
  • Replace some instances of Tensor::get_shape with get_logical_shape
  • Update METALIUM_GUIDE.md
  • #16621: Add barriers at end of cq_dispatch_slave.cpp on IERISC
  • Finish porting OPs to compute_output_specs
  • ScopedGraphCapture
  • #15756 Pull in BH LLK fix for maxpool hang
  • #15246: Add sweep tests for logical_and, logical_or, logical_xor
  • #0: (MINOR) Bump to v0.55.0
  • #11512: Add sweeps for eltwise sharded ops 3
  • Add sweeps for unary, unary_sharded and binary_sharded versions of ops: fmod, remainder, maximum, minimum.
  • Don't leak tt_cluster.hpp through kernel_types.hpp
  • #6983: Renable skipped TT-NN unit test
  • #15450: Remove default values from circular buffer parameters in LLK compute APIs
  • update build flag on programming examples docs
  • Fix for P100 board type
  • Sever TT-Train's dependency on TT-Metalium's tests
  • [TT-Train] Update generate of LLM
  • [TT-Train] Add bias=false in LinearLayer
  • TT-Fabric Bringup Initial Check-in
  • #0: Sanitize writes to mailbox on ethernet cores.
  • Add Llama11B-N300 and Llama70B-TG (TP=32) to LLM table in README.md
  • [skip ci] Update llms.md
  • Update test_slice.py
  • #16625: Refactor tracking of sub-device managers from Device to a new class
  • Update code-analysis.yaml
  • [skip ci] Update llms.md
  • remove references to LFS
  • Fixes for conversion to row major for 0D and 0-volume tensors
  • #0: Disable BH tools test at workflow level
  • Removing some usages of LegacyShape, improve Tensor::to_string
  • [skip ci] Fix lint on a doc
  • #0: API Unification for Device and MeshDevice
  • Port ttnn::random and uniform from LegacyShape to SimpleShape
  • #16379: make softmax call moreh_softmax if rank above 4
  • #7126: remove skip for test_sd_matmul test
  • #0: Make device an optional parameter in the tensor distribution API
  • Added build-wheels to fast-dispatch-build-and-unit-tests-wrapper.yaml
  • Adding CCL Async test cases to TG nightly and bug fix
  • #11119: Move op_profiler.hpp under the ttnn folder
  • #15979: Switch to google benchmark for pgm dispatch tests
  • [tt-train] Add weight tying option for NanoGPT demo
  • #0: Fix build of test_pgm_dispatch
  • [tt-train] Update serialization of tensor for DDP
  • #0: Fix failing TG regression tests
  • [skip ci] Update llms.md
  • Add tiled interleaved permute for when width dimension doesn't move (row-major tiled invariant)
  • Add Fabric Router Config to to Hal
  • [skip ci] Update llms.md
  • Reflect ARCH_NAME Changes in CI Workflows
  • [skip ci] Update llms.md