Update e2bqwen.py
Browse files- e2bqwen.py +8 -11
e2bqwen.py
CHANGED
|
@@ -475,8 +475,7 @@ class QwenVLAPIModel(Model):
|
|
| 475 |
max_tokens=4096,
|
| 476 |
)
|
| 477 |
self.fallback_model = HfApiModel(
|
| 478 |
-
model_id,
|
| 479 |
-
provider="nebius",
|
| 480 |
token=hf_token,
|
| 481 |
max_tokens=4096,
|
| 482 |
)
|
|
@@ -491,12 +490,10 @@ class QwenVLAPIModel(Model):
|
|
| 491 |
message = self.base_model(messages, stop_sequences, **kwargs)
|
| 492 |
return message
|
| 493 |
except Exception as e:
|
| 494 |
-
|
| 495 |
-
#
|
| 496 |
-
|
| 497 |
-
|
| 498 |
-
|
| 499 |
-
|
| 500 |
-
|
| 501 |
-
# except Exception as e:
|
| 502 |
-
# raise Exception(f"Both endpoints failed. Last error: {e}")
|
|
|
|
| 475 |
max_tokens=4096,
|
| 476 |
)
|
| 477 |
self.fallback_model = HfApiModel(
|
| 478 |
+
model_id="https://n5wr7lfx6wp94tvl.us-east-1.aws.endpoints.huggingface.cloud",
|
|
|
|
| 479 |
token=hf_token,
|
| 480 |
max_tokens=4096,
|
| 481 |
)
|
|
|
|
| 490 |
message = self.base_model(messages, stop_sequences, **kwargs)
|
| 491 |
return message
|
| 492 |
except Exception as e:
|
| 493 |
+
print(f"Base model failed with error: {e}. Calling fallback model.")
|
| 494 |
+
# Continue to fallback
|
| 495 |
+
try:
|
| 496 |
+
message = self.fallback_model(messages, stop_sequences, **kwargs)
|
| 497 |
+
return message
|
| 498 |
+
except Exception as e:
|
| 499 |
+
raise Exception(f"Both endpoints failed. Last error: {e}")
|
|
|
|
|
|