Working inference
Browse files- app.py +5 -5
- e2bqwen.py +11 -13
app.py
CHANGED
|
@@ -28,10 +28,6 @@ if not os.path.exists(TMP_DIR):
|
|
| 28 |
|
| 29 |
hf_token = os.getenv("HUGGINGFACE_API_KEY")
|
| 30 |
login(token=hf_token)
|
| 31 |
-
model = QwenVLAPIModel(
|
| 32 |
-
hf_base_url="https://s41ydkv0iyjeokyj.us-east-1.aws.endpoints.huggingface.cloud",
|
| 33 |
-
hf_token = hf_token,
|
| 34 |
-
)
|
| 35 |
|
| 36 |
|
| 37 |
custom_css = """
|
|
@@ -467,12 +463,16 @@ def update_terminal_from_session(session_hash):
|
|
| 467 |
|
| 468 |
|
| 469 |
def create_agent(data_dir, desktop, log_file):
|
|
|
|
|
|
|
|
|
|
|
|
|
| 470 |
return E2BVisionAgent(
|
| 471 |
model=model,
|
| 472 |
data_dir=data_dir,
|
| 473 |
desktop=desktop,
|
| 474 |
max_steps=200,
|
| 475 |
-
verbosity_level=
|
| 476 |
planning_interval=10,
|
| 477 |
log_file = log_file
|
| 478 |
)
|
|
|
|
| 28 |
|
| 29 |
hf_token = os.getenv("HUGGINGFACE_API_KEY")
|
| 30 |
login(token=hf_token)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
|
| 33 |
custom_css = """
|
|
|
|
| 463 |
|
| 464 |
|
| 465 |
def create_agent(data_dir, desktop, log_file):
|
| 466 |
+
model = QwenVLAPIModel(
|
| 467 |
+
model_id="Qwen/Qwen2.5-VL-72B-Instruct",
|
| 468 |
+
hf_token = hf_token,
|
| 469 |
+
)
|
| 470 |
return E2BVisionAgent(
|
| 471 |
model=model,
|
| 472 |
data_dir=data_dir,
|
| 473 |
desktop=desktop,
|
| 474 |
max_steps=200,
|
| 475 |
+
verbosity_level=2,
|
| 476 |
planning_interval=10,
|
| 477 |
log_file = log_file
|
| 478 |
)
|
e2bqwen.py
CHANGED
|
@@ -97,7 +97,7 @@ class E2BVisionAgent(CodeAgent):
|
|
| 97 |
desktop: Sandbox,
|
| 98 |
tools: List[tool] = None,
|
| 99 |
max_steps: int = 200,
|
| 100 |
-
verbosity_level: LogLevel =
|
| 101 |
planning_interval: int = 10,
|
| 102 |
log_file = None,
|
| 103 |
**kwargs
|
|
@@ -352,21 +352,19 @@ class QwenVLAPIModel(Model):
|
|
| 352 |
|
| 353 |
def __init__(
|
| 354 |
self,
|
| 355 |
-
|
| 356 |
-
model_path: str = "Qwen/Qwen2.5-VL-72B-Instruct",
|
| 357 |
-
provider: str = "hyperbolic",
|
| 358 |
hf_token: str = None,
|
| 359 |
):
|
| 360 |
super().__init__()
|
| 361 |
-
self.model_id =
|
| 362 |
-
self.
|
| 363 |
-
|
| 364 |
-
|
| 365 |
-
token=hf_token
|
| 366 |
)
|
| 367 |
self.fallback_model = HfApiModel(
|
| 368 |
-
|
| 369 |
-
provider=
|
| 370 |
token=hf_token,
|
| 371 |
)
|
| 372 |
|
|
@@ -378,9 +376,9 @@ class QwenVLAPIModel(Model):
|
|
| 378 |
) -> ChatMessage:
|
| 379 |
|
| 380 |
try:
|
| 381 |
-
return self.
|
| 382 |
except Exception as e:
|
| 383 |
-
print(f"
|
| 384 |
|
| 385 |
# Continue to fallback
|
| 386 |
try:
|
|
|
|
| 97 |
desktop: Sandbox,
|
| 98 |
tools: List[tool] = None,
|
| 99 |
max_steps: int = 200,
|
| 100 |
+
verbosity_level: LogLevel = 2,
|
| 101 |
planning_interval: int = 10,
|
| 102 |
log_file = None,
|
| 103 |
**kwargs
|
|
|
|
| 352 |
|
| 353 |
def __init__(
|
| 354 |
self,
|
| 355 |
+
model_id: str = "Qwen/Qwen2.5-VL-72B-Instruct",
|
|
|
|
|
|
|
| 356 |
hf_token: str = None,
|
| 357 |
):
|
| 358 |
super().__init__()
|
| 359 |
+
self.model_id = model_id
|
| 360 |
+
self.base_model = HfApiModel(
|
| 361 |
+
model_id,
|
| 362 |
+
provider="nebius",
|
| 363 |
+
token=hf_token,
|
| 364 |
)
|
| 365 |
self.fallback_model = HfApiModel(
|
| 366 |
+
model_id,
|
| 367 |
+
provider="hyperbolic",
|
| 368 |
token=hf_token,
|
| 369 |
)
|
| 370 |
|
|
|
|
| 376 |
) -> ChatMessage:
|
| 377 |
|
| 378 |
try:
|
| 379 |
+
return self.base_model(messages, stop_sequences, **kwargs)
|
| 380 |
except Exception as e:
|
| 381 |
+
print(f"Base model failed with error: {e}. Calling fallback model.")
|
| 382 |
|
| 383 |
# Continue to fallback
|
| 384 |
try:
|