ngebodh commited on
Commit
93447a2
·
verified ·
1 Parent(s): d879f4b

Llama 3.1 Serverless inference not available.

Browse files

Llama 3.1 Serverless inference not available.
Needs pro sub.

TODO: Fix with quantized version?

Error Message:
BadRequestError: Error code: 400 - {'error': 'Model requires a Pro subscription; check out hf.co/pricing to learn more. Make sure to include your HF token in your query.'}

Traceback:
File "/home/user/app/app.py", line 180, in <module>
stream = client.chat.completions.create(
File "/usr/local/lib/python3.10/site-packages/openai/_utils/_utils.py", line 277, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.10/site-packages/openai/resources/chat/completions.py", line 646, in create
return self._post(
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1266, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 942, in request
return self._request(
File "/usr/local/lib/python3.10/site-packages/openai/_base_client.py", line 1046, in _request
raise self._make_status_error_from_response(err.response) from None

Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -26,12 +26,12 @@ client = OpenAI(
26
 
27
  #Create supported models
28
  model_links ={
29
- "Meta-Llama-3.1-8B":"meta-llama/Meta-Llama-3.1-8B-Instruct",
30
  "Meta-Llama-3-8B":"meta-llama/Meta-Llama-3-8B-Instruct",
31
  "Mistral-7B":"mistralai/Mistral-7B-Instruct-v0.2",
32
  "Gemma-7B":"google/gemma-1.1-7b-it",
33
  "Gemma-2B":"google/gemma-1.1-2b-it",
34
  "Zephyr-7B-β":"HuggingFaceH4/zephyr-7b-beta",
 
35
 
36
  }
37
 
 
26
 
27
  #Create supported models
28
  model_links ={
 
29
  "Meta-Llama-3-8B":"meta-llama/Meta-Llama-3-8B-Instruct",
30
  "Mistral-7B":"mistralai/Mistral-7B-Instruct-v0.2",
31
  "Gemma-7B":"google/gemma-1.1-7b-it",
32
  "Gemma-2B":"google/gemma-1.1-2b-it",
33
  "Zephyr-7B-β":"HuggingFaceH4/zephyr-7b-beta",
34
+ #"Meta-Llama-3.1-8B":"meta-llama/Meta-Llama-3.1-8B-Instruct", #TODO: Update when/if Serverless Inference available
35
 
36
  }
37