roman-bachmann commited on
Commit
9199279
·
verified ·
1 Parent(s): bbce393

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -7
app.py CHANGED
@@ -16,11 +16,6 @@ from flextok.flextok_wrapper import FlexTokFromHub
16
  from flextok.utils.demo import imgs_from_urls, denormalize, batch_to_pil
17
  from flextok.utils.misc import detect_bf16_support, get_bf16_context, get_generator
18
 
19
- #import torch._dynamo
20
- #torch._dynamo.config.suppress_errors = True
21
- os.environ["TORCH_LOGS"] = "+dynamo"
22
- os.environ["TORCHDYNAMO_VERBOSE"] = "1"
23
-
24
  # We recommend running this demo on an A100 GPU
25
  if torch.cuda.is_available():
26
  device = "cuda"
@@ -53,8 +48,14 @@ MODEL_NAME = 'FlexTok d18-d28 (DFN)'
53
 
54
  # Load FlexTok model from HF Hub
55
  flextok_model = FlexTokFromHub.from_pretrained(MODEL_ID).to(device).eval()
56
- #flextok_model.encoder.module_dict.enc_seq_packer.compile_block_mask = False
57
- #flextok_model.decoder.module_dict.dec_seq_packer.compile_block_mask = False
 
 
 
 
 
 
58
 
59
  # Load AuraSR model from HF Hub
60
  try:
 
16
  from flextok.utils.demo import imgs_from_urls, denormalize, batch_to_pil
17
  from flextok.utils.misc import detect_bf16_support, get_bf16_context, get_generator
18
 
 
 
 
 
 
19
  # We recommend running this demo on an A100 GPU
20
  if torch.cuda.is_available():
21
  device = "cuda"
 
48
 
49
  # Load FlexTok model from HF Hub
50
  flextok_model = FlexTokFromHub.from_pretrained(MODEL_ID).to(device).eval()
51
+
52
+ # Disable flex_attention for HF Space
53
+ flextok_model.encoder.module_dict.enc_seq_packer.return_materialized_mask = True
54
+ flextok_model.decoder.module_dict.dec_seq_packer.return_materialized_mask = True
55
+ for block in flextok_model.encoder.module_dict.enc_transformer.blocks:
56
+ block._checkpoint_wrapped_module.attn.use_flex_attention = False
57
+ for block in flextok_model.decoder.module_dict.dec_transformer.blocks:
58
+ block._checkpoint_wrapped_module.attn.use_flex_attention = False
59
 
60
  # Load AuraSR model from HF Hub
61
  try: