first commit
Browse files- Dockerfile +20 -0
- README.md +2 -0
- config.yaml +24 -0
Dockerfile
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Use the provided base image
|
2 |
+
FROM ghcr.io/berriai/litellm:main-latest
|
3 |
+
|
4 |
+
# Set the working directory to /app
|
5 |
+
WORKDIR /app
|
6 |
+
|
7 |
+
# Copy the configuration file into the container at /app
|
8 |
+
COPY config.yaml .
|
9 |
+
|
10 |
+
# Make sure your entrypoint.sh is executable
|
11 |
+
RUN chmod +x entrypoint.sh
|
12 |
+
|
13 |
+
# Expose the necessary port
|
14 |
+
EXPOSE 4000/tcp
|
15 |
+
|
16 |
+
# Override the CMD instruction with your desired command and arguments
|
17 |
+
# WARNING: FOR PROD DO NOT USE `--detailed_debug` it slows down response times, instead use the following CMD
|
18 |
+
# CMD ["--port", "4000", "--config", "config.yaml"]
|
19 |
+
|
20 |
+
CMD ["--port", "4000", "--config", "config.yaml"]
|
README.md
CHANGED
@@ -5,6 +5,8 @@ colorFrom: blue
|
|
5 |
colorTo: purple
|
6 |
sdk: docker
|
7 |
pinned: false
|
|
|
|
|
8 |
---
|
9 |
|
10 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
5 |
colorTo: purple
|
6 |
sdk: docker
|
7 |
pinned: false
|
8 |
+
app_port: 4000
|
9 |
+
license: mit
|
10 |
---
|
11 |
|
12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
config.yaml
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model_list:
|
2 |
+
- model_name: gpt-4o
|
3 |
+
litellm_params:
|
4 |
+
model: github/gpt-4o
|
5 |
+
api_base: https://models.inference.ai.azure.com
|
6 |
+
api_key: "os.environ/GITHUB_API_KEY"
|
7 |
+
- model_name: gpt-4o-mini
|
8 |
+
litellm_params:
|
9 |
+
model: github/gpt-4o-mini
|
10 |
+
api_base: https://models.inference.ai.azure.com
|
11 |
+
api_key: "os.environ/GITHUB_API_KEY"
|
12 |
+
- model_name: meta-llama-3.1-405b-instruct
|
13 |
+
litellm_params:
|
14 |
+
model: github/meta-llama-3.1-405b-instruct
|
15 |
+
api_base: https://models.inference.ai.azure.com
|
16 |
+
api_key: "os.environ/GITHUB_API_KEY"
|
17 |
+
- model_name: meta-llama-3.1-8b-instruct
|
18 |
+
litellm_params:
|
19 |
+
model: github/meta-llama-3.1-8b-instruct
|
20 |
+
api_base: https://models.inference.ai.azure.com
|
21 |
+
api_key: "os.environ/GITHUB_API_KEY"
|
22 |
+
|
23 |
+
litellm_settings:
|
24 |
+
drop_params: True
|