Skip to content

Commit e6f05d1

Browse files
authored
update release notes (#1502)
1 parent 9878e4f commit e6f05d1

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

docs/tutorials/releases.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@ We are pleased to announce the release of Intel® Extension for PyTorch\* 2.0.0-
99

1010
- **Fast BERT optimization (Experimental)**: Intel introduced a new technique to speed up BERT workloads. Intel® Extension for PyTorch\* integrated this implementation, which benefits BERT model especially training. A new API `ipex.fast_bert` is provided to try this new optimization. More detailed information can be found at [Fast Bert Feature](./features/fast_bert.md).
1111

12+
- **MHA optimization with Flash Attention**: Intel optimized MHA module with Flash Attention technique as inspired by [Stanford paper](https://arxiv.org/abs/2205.14135). This brings less memory consumption for LLM, and also provides better inference performance for models like BERT, Stable Diffusion, etc.
13+
1214
- **Work with torch.compile as an backend (Experimental)**: PyTorch 2.0 introduces a new feature, `torch.compile`, to speed up PyTorch execution. We've enabled Intel® Extension for PyTorch as a backend of torch.compile, which can leverage this new PyTorch API's power of graph capture and provide additional optimization based on these graphs.
1315
The usage of this new feature is quite simple as below:
1416

0 commit comments

Comments
 (0)