Spaces:
Running
Llama 3.1 Serverless inference not available.
Browse filesLlama 3.1 Serverless inference not available.
Needs pro sub.
TODO: Fix with quantized version?
Error Message:
BadRequestError: Error code: 400 - {'error': 'Model requires a Pro subscription; check out hf.co/pricing to learn more. Make sure to include your HF token in your query.'}
Traceback:
File "/home/user/app/app.py", line 180, in <module>
stream = client.chat.completions.create(
File "/usr/local/lib/python3.10/site-packages/openai/_utils/_utils.py", line 277, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.10/site-packages/openai/resources/chat/completions.py", line 646, in create
return self._post(
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1266, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 942, in request
return self._request(
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1046, in _request
raise self._make_status_error_from_response(err.response) from None
@@ -26,12 +26,12 @@ client = OpenAI(
|
|
26 |
|
27 |
#Create supported models
|
28 |
model_links ={
|
29 |
-
"Meta-Llama-3.1-8B":"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
30 |
"Meta-Llama-3-8B":"meta-llama/Meta-Llama-3-8B-Instruct",
|
31 |
"Mistral-7B":"mistralai/Mistral-7B-Instruct-v0.2",
|
32 |
"Gemma-7B":"google/gemma-1.1-7b-it",
|
33 |
"Gemma-2B":"google/gemma-1.1-2b-it",
|
34 |
"Zephyr-7B-β":"HuggingFaceH4/zephyr-7b-beta",
|
|
|
35 |
|
36 |
}
|
37 |
|
|
|
26 |
|
27 |
#Create supported models
|
28 |
model_links ={
|
|
|
29 |
"Meta-Llama-3-8B":"meta-llama/Meta-Llama-3-8B-Instruct",
|
30 |
"Mistral-7B":"mistralai/Mistral-7B-Instruct-v0.2",
|
31 |
"Gemma-7B":"google/gemma-1.1-7b-it",
|
32 |
"Gemma-2B":"google/gemma-1.1-2b-it",
|
33 |
"Zephyr-7B-β":"HuggingFaceH4/zephyr-7b-beta",
|
34 |
+
#"Meta-Llama-3.1-8B":"meta-llama/Meta-Llama-3.1-8B-Instruct", #TODO: Update when/if Serverless Inference available
|
35 |
|
36 |
}
|
37 |
|