Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Working inference
Browse files- app.py +5 -5
- e2bqwen.py +11 -13
app.py
CHANGED
@@ -28,10 +28,6 @@ if not os.path.exists(TMP_DIR):
|
|
28 |
|
29 |
hf_token = os.getenv("HUGGINGFACE_API_KEY")
|
30 |
login(token=hf_token)
|
31 |
-
model = QwenVLAPIModel(
|
32 |
-
hf_base_url="https://s41ydkv0iyjeokyj.us-east-1.aws.endpoints.huggingface.cloud",
|
33 |
-
hf_token = hf_token,
|
34 |
-
)
|
35 |
|
36 |
|
37 |
custom_css = """
|
@@ -467,12 +463,16 @@ def update_terminal_from_session(session_hash):
|
|
467 |
|
468 |
|
469 |
def create_agent(data_dir, desktop, log_file):
|
|
|
|
|
|
|
|
|
470 |
return E2BVisionAgent(
|
471 |
model=model,
|
472 |
data_dir=data_dir,
|
473 |
desktop=desktop,
|
474 |
max_steps=200,
|
475 |
-
verbosity_level=
|
476 |
planning_interval=10,
|
477 |
log_file = log_file
|
478 |
)
|
|
|
28 |
|
29 |
hf_token = os.getenv("HUGGINGFACE_API_KEY")
|
30 |
login(token=hf_token)
|
|
|
|
|
|
|
|
|
31 |
|
32 |
|
33 |
custom_css = """
|
|
|
463 |
|
464 |
|
465 |
def create_agent(data_dir, desktop, log_file):
|
466 |
+
model = QwenVLAPIModel(
|
467 |
+
model_id="Qwen/Qwen2.5-VL-72B-Instruct",
|
468 |
+
hf_token = hf_token,
|
469 |
+
)
|
470 |
return E2BVisionAgent(
|
471 |
model=model,
|
472 |
data_dir=data_dir,
|
473 |
desktop=desktop,
|
474 |
max_steps=200,
|
475 |
+
verbosity_level=2,
|
476 |
planning_interval=10,
|
477 |
log_file = log_file
|
478 |
)
|
e2bqwen.py
CHANGED
@@ -97,7 +97,7 @@ class E2BVisionAgent(CodeAgent):
|
|
97 |
desktop: Sandbox,
|
98 |
tools: List[tool] = None,
|
99 |
max_steps: int = 200,
|
100 |
-
verbosity_level: LogLevel =
|
101 |
planning_interval: int = 10,
|
102 |
log_file = None,
|
103 |
**kwargs
|
@@ -352,21 +352,19 @@ class QwenVLAPIModel(Model):
|
|
352 |
|
353 |
def __init__(
|
354 |
self,
|
355 |
-
|
356 |
-
model_path: str = "Qwen/Qwen2.5-VL-72B-Instruct",
|
357 |
-
provider: str = "hyperbolic",
|
358 |
hf_token: str = None,
|
359 |
):
|
360 |
super().__init__()
|
361 |
-
self.model_id =
|
362 |
-
self.
|
363 |
-
|
364 |
-
|
365 |
-
token=hf_token
|
366 |
)
|
367 |
self.fallback_model = HfApiModel(
|
368 |
-
|
369 |
-
provider=
|
370 |
token=hf_token,
|
371 |
)
|
372 |
|
@@ -378,9 +376,9 @@ class QwenVLAPIModel(Model):
|
|
378 |
) -> ChatMessage:
|
379 |
|
380 |
try:
|
381 |
-
return self.
|
382 |
except Exception as e:
|
383 |
-
print(f"
|
384 |
|
385 |
# Continue to fallback
|
386 |
try:
|
|
|
97 |
desktop: Sandbox,
|
98 |
tools: List[tool] = None,
|
99 |
max_steps: int = 200,
|
100 |
+
verbosity_level: LogLevel = 2,
|
101 |
planning_interval: int = 10,
|
102 |
log_file = None,
|
103 |
**kwargs
|
|
|
352 |
|
353 |
def __init__(
|
354 |
self,
|
355 |
+
model_id: str = "Qwen/Qwen2.5-VL-72B-Instruct",
|
|
|
|
|
356 |
hf_token: str = None,
|
357 |
):
|
358 |
super().__init__()
|
359 |
+
self.model_id = model_id
|
360 |
+
self.base_model = HfApiModel(
|
361 |
+
model_id,
|
362 |
+
provider="nebius",
|
363 |
+
token=hf_token,
|
364 |
)
|
365 |
self.fallback_model = HfApiModel(
|
366 |
+
model_id,
|
367 |
+
provider="hyperbolic",
|
368 |
token=hf_token,
|
369 |
)
|
370 |
|
|
|
376 |
) -> ChatMessage:
|
377 |
|
378 |
try:
|
379 |
+
return self.base_model(messages, stop_sequences, **kwargs)
|
380 |
except Exception as e:
|
381 |
+
print(f"Base model failed with error: {e}. Calling fallback model.")
|
382 |
|
383 |
# Continue to fallback
|
384 |
try:
|