ankandrew
commited on
Commit
·
e8a0cf2
1
Parent(s):
f17ef4c
Try different flash attention version
Browse files
app.py
CHANGED
@@ -10,7 +10,7 @@ from qwen_vl_utils import process_vision_info
|
|
10 |
from transformers.utils import is_flash_attn_2_available
|
11 |
|
12 |
subprocess.run(
|
13 |
-
"pip install 'flash-attn
|
14 |
env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
|
15 |
shell=True,
|
16 |
)
|
|
|
10 |
from transformers.utils import is_flash_attn_2_available
|
11 |
|
12 |
subprocess.run(
|
13 |
+
"pip install 'flash-attn==2.2.0' --no-build-isolation",
|
14 |
env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
|
15 |
shell=True,
|
16 |
)
|