Skip to content

Releases: amd/ZenDNN-pytorch-plugin

zentorch Release v5.0

15 Nov 13:05
Compare
Choose a tag to compare

zentorch is compatible with base versions of PyTorch v2.0 or later. This release provides zentorch for PyTorch v2.4.0.
This release of the plug-in supports:

  • Datatypes FP32, BF16, and INT4 (WOQ)
  • Introduction of a new zentorch.llm.optimize() method for Hugging Face Generative LLM models
  • New zentorch.load_woq_model() method to support loading of Weight Only Quantized models generated through the AMD Quark tool. This method only supports models quantized and exported with per-channel quantization using the AWQ algorithm.
  • Improved graph optimizations, enhanced SDPA (Scalar Dot Product Attention) operator and more.
  • Automatic Mixed Precision (AMP) between FP32 and BF16 providing a performance improvement with minimal changes in accuracy

zentorch Release v4.2

22 May 14:55
Compare
Choose a tag to compare

This zentorch release:

  • Is compatible with PyTorch v2.0 and later
  • Extends Pytorch by providing a custom backend to torch.compile flow to provide a performant AI inference solution for AMD EPYCTM servers leveraging the ZenDNN 4.2 library
  • Includes graph optimizations and fusions tailored for AMD EPYCTM architectures
  • Supports BF16 execution through auto-mixed precision to provide performance improvements with minimal changes in accuracy