From 4ef63100463977f3aaddb22815a507f816505048 Mon Sep 17 00:00:00 2001 From: Antoni Baum Date: Wed, 12 Jun 2024 16:43:27 -0700 Subject: [PATCH] Pin vllm-flash-attn==v2.5.9.post1 --- requirements-cuda.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/requirements-cuda.txt b/requirements-cuda.txt index 353617983596..5eb4075874b0 100644 --- a/requirements-cuda.txt +++ b/requirements-cuda.txt @@ -6,4 +6,4 @@ ray >= 2.9 nvidia-ml-py # for pynvml package torch == 2.3.0 xformers == 0.0.26.post1 # Requires PyTorch 2.3.0 -vllm-flash-attn == 2.5.9 # Requires PyTorch 2.3.0 +vllm-flash-attn == v2.5.9.post1 # Requires PyTorch 2.3.0