Sofia Casadei commited on
Commit
87d5645
·
1 Parent(s): 157b1a9

install flash attention

Browse files
Files changed (1) hide show
  1. Dockerfile +2 -3
Dockerfile CHANGED
@@ -45,10 +45,9 @@ COPY --chown=user requirements.txt .
45
  RUN --mount=type=cache,target=$UV_CACHE_DIR,uid=1000,gid=1000 \
46
  /uv pip install -r requirements.txt
47
 
48
- # Try to install flash-attn, but don't fail the build if it errors
49
  RUN --mount=type=cache,target=$UV_CACHE_DIR,uid=1000,gid=1000 \
50
- echo "Attempting to install flash-attn..." && \
51
- (/uv pip install flash-attn --no-build-isolation || echo "Failed, continuing build...")
52
 
53
  # Copy application code
54
  COPY --chown=user . .
 
45
  RUN --mount=type=cache,target=$UV_CACHE_DIR,uid=1000,gid=1000 \
46
  /uv pip install -r requirements.txt
47
 
48
+ # Install flash-attn from pre-built wheel
49
  RUN --mount=type=cache,target=$UV_CACHE_DIR,uid=1000,gid=1000 \
50
+ /uv pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.6cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
 
51
 
52
  # Copy application code
53
  COPY --chown=user . .