From 9878e4f2834dd1db880fddd2404489d4d00f3161 Mon Sep 17 00:00:00 2001 From: Jing Xu Date: Wed, 22 Mar 2023 14:07:30 +0900 Subject: [PATCH] update docs for FastBert (#1496) --- docs/tutorials/features.rst | 4 +--- docs/tutorials/features/fast_bert.md | 4 ++-- docs/tutorials/releases.md | 2 +- 3 files changed, 4 insertions(+), 6 deletions(-) diff --git a/docs/tutorials/features.rst b/docs/tutorials/features.rst index 62571af64..0261f6213 100644 --- a/docs/tutorials/features.rst +++ b/docs/tutorials/features.rst @@ -201,9 +201,7 @@ For more detailed information, check `HyperTune `_. Fast BERT Optimization (Experimental, *NEW feature from 2.0.0*) --------------------------------------------------------------- -Intel proposed a technique, Tensor Processing Primitives (TPP), a programming abstraction striving for efficient, portable implementation of DL workloads with high-productivity. TPPs define a compact, yet versatile set of 2D-tensor operators (or a virtual Tensor ISA), which subsequently can be utilized as building-blocks to construct complex operators on high-dimensional tensors. - -Implementation of TPP is integrated into Intel® Extension for PyTorch\*. BERT could benefit from this new technique. An API `ipex.fast_bert` is provided for a simple usage. +Intel proposed a technique to speed up BERT workloads. Implementation is integrated into Intel® Extension for PyTorch\*. An API `ipex.fast_bert` is provided for a simple usage. For more detailed information, check `Fast BERT `_. diff --git a/docs/tutorials/features/fast_bert.md b/docs/tutorials/features/fast_bert.md index 08e121063..f84bbcc74 100644 --- a/docs/tutorials/features/fast_bert.md +++ b/docs/tutorials/features/fast_bert.md @@ -3,9 +3,9 @@ Fast BERT (Experimental) ### Feature Description -Intel proposed a technique, Tensor Processing Primitives (TPP), a programming abstraction striving for efficient, portable implementation of DL workloads with high-productivity. TPPs define a compact, yet versatile set of 2D-tensor operators (or a virtual Tensor ISA), which subsequently can be utilized as building-blocks to construct complex operators on high-dimensional tensors. Detailed contents are available at [*Tensor Processing Primitives: A Programming Abstraction for Efficiency and Portability in Deep Learning & HPC Workloads*](https://arxiv.org/pdf/2104.05755.pdf). +Intel proposed a technique to speed up BERT workloads. Implementation leverages the idea from [*Tensor Processing Primitives: A Programming Abstraction for Efficiency and Portability in Deep Learning & HPC Workloads*](https://arxiv.org/pdf/2104.05755.pdf). -Implementation of TPP is integrated into Intel® Extension for PyTorch\*. BERT could benefit from this new technique, for both training and inference. +The Implementation is integrated into Intel® Extension for PyTorch\*. BERT could benefit from this new technique, for both training and inference. ### Prerequisite diff --git a/docs/tutorials/releases.md b/docs/tutorials/releases.md index e7d9afbc5..c911a4003 100644 --- a/docs/tutorials/releases.md +++ b/docs/tutorials/releases.md @@ -7,7 +7,7 @@ We are pleased to announce the release of Intel® Extension for PyTorch\* 2.0.0- ### Highlights -- **Fast BERT optimization (Experimental)**: Intel introduced a new technique, Tensor Processing Primitives (TPP), a programming abstraction striving for efficient, portable implementation of DL workloads with high-productivity. Intel® Extension for PyTorch\* integrated this TPP implementation, which benefits BERT model especially training. A new API `ipex.fast_bert` is provided to try this new optimization. More detailed information can be found at [Fast Bert Feature](./features/fast_bert.md). +- **Fast BERT optimization (Experimental)**: Intel introduced a new technique to speed up BERT workloads. Intel® Extension for PyTorch\* integrated this implementation, which benefits BERT model especially training. A new API `ipex.fast_bert` is provided to try this new optimization. More detailed information can be found at [Fast Bert Feature](./features/fast_bert.md). - **Work with torch.compile as an backend (Experimental)**: PyTorch 2.0 introduces a new feature, `torch.compile`, to speed up PyTorch execution. We've enabled Intel® Extension for PyTorch as a backend of torch.compile, which can leverage this new PyTorch API's power of graph capture and provide additional optimization based on these graphs. The usage of this new feature is quite simple as below: