From 280cb6c2ecbe3813b9b87e7b722d481fcc276db5 Mon Sep 17 00:00:00 2001 From: zzhhjjj Date: Thu, 2 May 2024 11:59:56 +0000 Subject: [PATCH] add flash attention --- .github/workflows/llama_tests.yaml | 1 + 1 file changed, 1 insertion(+) diff --git a/.github/workflows/llama_tests.yaml b/.github/workflows/llama_tests.yaml index eee801c7..61ac7d79 100644 --- a/.github/workflows/llama_tests.yaml +++ b/.github/workflows/llama_tests.yaml @@ -43,6 +43,7 @@ jobs: python -m pip install --upgrade pip pip install packaging pip install wheel + pip install "flash-attn>=2.5.0" --no-build-isolation pip install -e . pip install -e .[dev] pip install -e .[test]