Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -16,11 +16,6 @@ from flextok.flextok_wrapper import FlexTokFromHub
|
|
16 |
from flextok.utils.demo import imgs_from_urls, denormalize, batch_to_pil
|
17 |
from flextok.utils.misc import detect_bf16_support, get_bf16_context, get_generator
|
18 |
|
19 |
-
#import torch._dynamo
|
20 |
-
#torch._dynamo.config.suppress_errors = True
|
21 |
-
os.environ["TORCH_LOGS"] = "+dynamo"
|
22 |
-
os.environ["TORCHDYNAMO_VERBOSE"] = "1"
|
23 |
-
|
24 |
# We recommend running this demo on an A100 GPU
|
25 |
if torch.cuda.is_available():
|
26 |
device = "cuda"
|
@@ -53,8 +48,14 @@ MODEL_NAME = 'FlexTok d18-d28 (DFN)'
|
|
53 |
|
54 |
# Load FlexTok model from HF Hub
|
55 |
flextok_model = FlexTokFromHub.from_pretrained(MODEL_ID).to(device).eval()
|
56 |
-
|
57 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
58 |
|
59 |
# Load AuraSR model from HF Hub
|
60 |
try:
|
|
|
16 |
from flextok.utils.demo import imgs_from_urls, denormalize, batch_to_pil
|
17 |
from flextok.utils.misc import detect_bf16_support, get_bf16_context, get_generator
|
18 |
|
|
|
|
|
|
|
|
|
|
|
19 |
# We recommend running this demo on an A100 GPU
|
20 |
if torch.cuda.is_available():
|
21 |
device = "cuda"
|
|
|
48 |
|
49 |
# Load FlexTok model from HF Hub
|
50 |
flextok_model = FlexTokFromHub.from_pretrained(MODEL_ID).to(device).eval()
|
51 |
+
|
52 |
+
# Disable flex_attention for HF Space
|
53 |
+
flextok_model.encoder.module_dict.enc_seq_packer.return_materialized_mask = True
|
54 |
+
flextok_model.decoder.module_dict.dec_seq_packer.return_materialized_mask = True
|
55 |
+
for block in flextok_model.encoder.module_dict.enc_transformer.blocks:
|
56 |
+
block._checkpoint_wrapped_module.attn.use_flex_attention = False
|
57 |
+
for block in flextok_model.decoder.module_dict.dec_transformer.blocks:
|
58 |
+
block._checkpoint_wrapped_module.attn.use_flex_attention = False
|
59 |
|
60 |
# Load AuraSR model from HF Hub
|
61 |
try:
|