From 1bc66896ecbab76ab94e6961c6ad6be6ffd39e30 Mon Sep 17 00:00:00 2001 From: JegernOUTT Date: Mon, 6 Nov 2023 19:26:15 +1030 Subject: [PATCH] do not install flash-attention by default --- Dockerfile | 1 + setup.py | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index 54b1a74f..917d4606 100644 --- a/Dockerfile +++ b/Dockerfile @@ -39,6 +39,7 @@ RUN git clone https://github.com/smallcloudai/linguist.git /tmp/linguist \ && rake build_gem ENV PATH="${PATH}:/tmp/linguist/bin" +ENV INSTALL_OPTIONAL=TRUE ENV BUILD_CUDA_EXT=1 ENV GITHUB_ACTIONS=true ENV TORCH_CUDA_ARCH_LIST="6.0;6.1;7.0;7.5;8.0;8.6;8.9;9.0+PTX" diff --git a/setup.py b/setup.py index 7b122bfe..7368e07c 100644 --- a/setup.py +++ b/setup.py @@ -8,7 +8,7 @@ from typing import List, Set setup_package = os.environ.get("SETUP_PACKAGE", None) -install_optional = os.environ.get("INSTALL_OPTIONAL", "TRUE") +install_optional = os.environ.get("INSTALL_OPTIONAL", "FALSE") # Setting some env variables to force flash-attention build from sources # We can get rid of them when https://github.com/Dao-AILab/flash-attention/pull/540 is merged