metadata
license: apache-2.0
tags:
- flash-attention
- nvidia
- blackwell
- windows
- prebuilt-wheels
- python
- machine-learning
- deep-learning
Flash-Attention 2.7.4 Prebuilt Wheels for NVIDIA Blackwell (cu128) on Windows
This repository provides prebuilt wheels for Flash-Attention 2.7.4 optimized for NVIDIA Blackwell GPUs (cu128) on Windows systems. These wheels are compatible with Python 3.10 and 3.11, enabling seamless integration for high-performance attention mechanisms in deep learning workflows.
Available Wheels
flash_attn-2.7.4.post1-cp310-cp310-win_amd64.whl
(Python 3.10)flash_attn-2.7.4.post1-cp311-cp311-win_amd64.whl
(Python 3.11)
Compatibility
The prebuilt wheels are designed for NVIDIA Blackwell GPUs but have been tested and confirmed compatible with previous-generation NVIDIA GPUs, including:
- NVIDIA RTX 5090
- NVIDIA RTX 3090
Installation
To install, use pip with the appropriate wheel for your Python version:
pip install flash_attn-2.7.4.post1-cp310-cp310-win_amd64.whl
# or
pip install flash_attn-2.7.4.post1-cp311-cp311-win_amd64.whl