Skip to content

Commit

Permalink
Prepare release 0.0.25 (fairinternal/xformers#1058)
Browse files Browse the repository at this point in the history
* Prepare release 0.0.25

* Add note about supported PT version for binaries

__original_commit__ = fairinternal/xformers@0510e14
  • Loading branch information
danthe3rd authored and xFormers Bot committed Mar 14, 2024
1 parent 503a5d7 commit 78f5d2e
Show file tree
Hide file tree
Showing 4 changed files with 16 additions and 10 deletions.
4 changes: 2 additions & 2 deletions .github/workflows/conda.yml
Original file line number Diff line number Diff line change
Expand Up @@ -33,13 +33,13 @@ jobs:
- "3.9"
- "3.10"
config:
- torch_version: "2.2.0"
- torch_version: "2.2.1"
torch_channel: "pytorch"
cuda_version: "12.1.0"
cuda_dep_runtime: ">=12.0,<13.0"
cuda_short_version: "121"

- torch_version: "2.2.0"
- torch_version: "2.2.1"
torch_channel: "pytorch"
cuda_version: "11.8.0"
cuda_dep_runtime: ">=11.7,<11.9"
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/wheels.yml
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ jobs:
- "3.10"
- "3.11"
torch_version:
- "2.2.0"
- "2.2.1"
cuda_short_version:
- "118"
- "121"
Expand All @@ -45,7 +45,7 @@ jobs:
uses: ./.github/workflows/wheels_upload_pip.yml
with:
twine_username: __token__
filter: "*torch2.2.0+cu121*"
filter: "*torch2.2.1+cu121*"
execute: ${{ github.repository == 'facebookresearch/xformers' && github.event_name != 'pull_request' }}
secrets:
twine_password: ${{ secrets.PYPI_TOKEN }}
Expand All @@ -57,7 +57,7 @@ jobs:
aws_role: "arn:aws:iam::749337293305:role/pytorch_bot_uploader_role"
s3_path: s3://pytorch/whl/cu118/
aws_s3_cp_extra_args: --acl public-read
filter: "*torch2.2.0+cu118*"
filter: "*torch2.2.1+cu118*"
execute: ${{ github.repository == 'facebookresearch/xformers' && github.ref_type == 'tag' }}

upload_pt_cu121:
Expand All @@ -67,6 +67,6 @@ jobs:
aws_role: "arn:aws:iam::749337293305:role/pytorch_bot_uploader_role"
s3_path: s3://pytorch/whl/cu121/
aws_s3_cp_extra_args: --acl public-read
filter: "*torch2.2.0+cu121*"
filter: "*torch2.2.1+cu121*"
execute: ${{ github.repository == 'facebookresearch/xformers' && github.ref_type == 'tag' }}

12 changes: 9 additions & 3 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,15 @@ All notable changes to this project will be documented in this file.
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).

## [0.0.25] - TBD
## [0.0.26] - TBD
### Added
- New merge_attentions function
### Improved
### Removed

## [0.0.25] - 2024-03-14
Pre-built binary wheels require PyTorch 2.2.1
### Added
- New `merge_attentions` function
### Improved
- fMHA: Updated Flash-Attention to v2.5.6: this has a performance improvement for multiquery.
- fMHA: triton_splitk changed and expanded. Now amalgamates using LSE. Can autotune, supports causal with a small number of queries - not just 1. Experimental support for paged attention.
Expand All @@ -18,7 +24,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
## [0.0.24] - 2024-01-31
Pre-built binary wheels require PyTorch 2.2.0
### Added
- Added components for model/sequence parallelism, as near-drop-in replacements for FairScale/Megatron Column&RowParallelLinear modules. They support fusing communication and computation for sequence parallelism, thus making the communication effectively free.
- Added components for model/sequence parallelism, as near-drop-in replacements for FairScale/Megatron Column&RowParallelLinear modules. They support fusing communication and computation for sequence parallelism, thus making the communication effectively free. [Read more](https://twitter.com/d_haziza/status/1753030654118211593)
- Added kernels for training models with 2:4-sparsity. We introduced a very fast kernel for converting a matrix A into 24-sparse format, which can be used during training to sparsify weights dynamically, activations etc... xFormers also provides an API that is compatible with torch-compile, see `xformers.ops.sparsify24`.
### Improved
- Make selective activation checkpointing be compatible with torch.compile.
Expand Down
2 changes: 1 addition & 1 deletion version.txt
Original file line number Diff line number Diff line change
@@ -1 +1 @@
0.0.25
0.0.26

0 comments on commit 78f5d2e

Please sign in to comment.