Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -89,7 +89,7 @@ def reset_components():
|
|
89 |
|
90 |
|
91 |
@spaces.GPU
|
92 |
-
def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
93 |
history = []
|
94 |
|
95 |
if not video:
|
@@ -115,6 +115,8 @@ def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
|
115 |
history.append({'role': 'user', 'content': prompt})
|
116 |
yield history
|
117 |
|
|
|
|
|
118 |
duration = get_duration(video)
|
119 |
|
120 |
# do grounding and answering by default
|
@@ -159,9 +161,7 @@ def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
|
159 |
|
160 |
images, videos = process_vision_info(messages)
|
161 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
162 |
-
data = data.to(
|
163 |
-
|
164 |
-
print(next(model.parameters()).device)
|
165 |
|
166 |
model.base_model.disable_adapter_layers()
|
167 |
model.base_model.enable_adapter_layers()
|
@@ -260,7 +260,7 @@ def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
|
260 |
text = processor.apply_chat_template(messages, add_generation_prompt=True)
|
261 |
images, videos = process_vision_info(messages)
|
262 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
263 |
-
data = data.to(
|
264 |
|
265 |
model.base_model.disable_adapter_layers()
|
266 |
model.base_model.enable_adapter_layers()
|
@@ -397,7 +397,7 @@ def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
|
397 |
data['attention_mask'] = torch.ones_like(data['input_ids'])
|
398 |
# ===========================================
|
399 |
|
400 |
-
data = data.to(
|
401 |
|
402 |
model.base_model.disable_adapter_layers()
|
403 |
model.base_model.enable_adapter_layers()
|
@@ -484,7 +484,7 @@ def main(video, prompt, role, temperature, max_new_tokens, model, processor):
|
|
484 |
text = processor.apply_chat_template(messages, add_generation_prompt=True)
|
485 |
images, videos = process_vision_info(messages)
|
486 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
487 |
-
data = data.to(
|
488 |
|
489 |
with model.disable_adapter():
|
490 |
output_ids = model.generate(
|
@@ -541,7 +541,9 @@ if __name__ == '__main__':
|
|
541 |
print('Initializing role *verifier*')
|
542 |
model.load_adapter(nncore.join(MODEL, 'verifier'), adapter_name='verifier')
|
543 |
|
544 |
-
|
|
|
|
|
545 |
|
546 |
path = os.path.dirname(os.path.realpath(__file__))
|
547 |
|
|
|
89 |
|
90 |
|
91 |
@spaces.GPU
|
92 |
+
def main(video, prompt, role, temperature, max_new_tokens, model, processor, device):
|
93 |
history = []
|
94 |
|
95 |
if not video:
|
|
|
115 |
history.append({'role': 'user', 'content': prompt})
|
116 |
yield history
|
117 |
|
118 |
+
model = model.to(device)
|
119 |
+
|
120 |
duration = get_duration(video)
|
121 |
|
122 |
# do grounding and answering by default
|
|
|
161 |
|
162 |
images, videos = process_vision_info(messages)
|
163 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
164 |
+
data = data.to(device)
|
|
|
|
|
165 |
|
166 |
model.base_model.disable_adapter_layers()
|
167 |
model.base_model.enable_adapter_layers()
|
|
|
260 |
text = processor.apply_chat_template(messages, add_generation_prompt=True)
|
261 |
images, videos = process_vision_info(messages)
|
262 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
263 |
+
data = data.to(device)
|
264 |
|
265 |
model.base_model.disable_adapter_layers()
|
266 |
model.base_model.enable_adapter_layers()
|
|
|
397 |
data['attention_mask'] = torch.ones_like(data['input_ids'])
|
398 |
# ===========================================
|
399 |
|
400 |
+
data = data.to(device)
|
401 |
|
402 |
model.base_model.disable_adapter_layers()
|
403 |
model.base_model.enable_adapter_layers()
|
|
|
484 |
text = processor.apply_chat_template(messages, add_generation_prompt=True)
|
485 |
images, videos = process_vision_info(messages)
|
486 |
data = processor(text=[text], images=images, videos=videos, return_tensors='pt')
|
487 |
+
data = data.to(device)
|
488 |
|
489 |
with model.disable_adapter():
|
490 |
output_ids = model.generate(
|
|
|
541 |
print('Initializing role *verifier*')
|
542 |
model.load_adapter(nncore.join(MODEL, 'verifier'), adapter_name='verifier')
|
543 |
|
544 |
+
device = torch.device('cuda')
|
545 |
+
|
546 |
+
main = partial(main, model=model, processor=processor, device=device)
|
547 |
|
548 |
path = os.path.dirname(os.path.realpath(__file__))
|
549 |
|