Spaces:
Paused
Paused
:pencil: [Doc] Update available models
Browse files
README.md
CHANGED
|
@@ -12,24 +12,15 @@ Huggingface LLM Inference API in OpenAI message format.
|
|
| 12 |
|
| 13 |
## Features
|
| 14 |
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
- Available Models:
|
| 18 |
-
- `mixtral-8x7b`, `mistral-7b`, `openchat-3.5`
|
| 19 |
- Adaptive prompt templates for different models
|
| 20 |
- Support OpenAI API format
|
| 21 |
-
-
|
| 22 |
- Support both stream and no-stream response
|
| 23 |
- Support API Key via both HTTP auth header and env varible (https://github.com/Hansimov/hf-llm-api/issues/4)
|
| 24 |
- Docker deployment
|
| 25 |
|
| 26 |
-
🔨 In progress:
|
| 27 |
-
- [ ] Support more models (https://github.com/Hansimov/hf-llm-api/issues/5)
|
| 28 |
-
- [ ] meta-llama/Llama-2-70b-chat-hf
|
| 29 |
-
- [ ] codellama/CodeLlama-34b-Instruct-hf
|
| 30 |
-
- [ ] tiiuae/falcon-180B-chat
|
| 31 |
-
|
| 32 |
-
|
| 33 |
## Run API service
|
| 34 |
|
| 35 |
### Run in Command Line
|
|
|
|
| 12 |
|
| 13 |
## Features
|
| 14 |
|
| 15 |
+
- Available Models (2024/01/15): (https://github.com/Hansimov/hf-llm-api/issues/5)
|
| 16 |
+
- `mixtral-8x7b`, `mistral-7b`
|
|
|
|
|
|
|
| 17 |
- Adaptive prompt templates for different models
|
| 18 |
- Support OpenAI API format
|
| 19 |
+
- Enable api endpoint via official `openai-python` package
|
| 20 |
- Support both stream and no-stream response
|
| 21 |
- Support API Key via both HTTP auth header and env varible (https://github.com/Hansimov/hf-llm-api/issues/4)
|
| 22 |
- Docker deployment
|
| 23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
## Run API service
|
| 25 |
|
| 26 |
### Run in Command Line
|