PyTorch 2.9 introduces stable libtorch ABI for C++/CUDA extensions, symmetric memory for multi-GPU kernel programming, and flexible graph break handling in torch.compile. The release expands hardware support with ROCm, XPU, and CUDA 13 wheel variants, enables FlexAttention on Intel GPUs, and adds flash decoding optimization for

6m read timeFrom pytorch.org
Post cover image
Table of contents
API-UNSTABLE FEATURES

Sort: