Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
β’
95ca61b
1
Parent(s):
3db3a54
more clarifications
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ You have three options:
|
|
40 |
|
41 |
### Option 1: Use an Inference API model
|
42 |
|
43 |
-
This is a new option added recently, where you can use one of the models from the Hugging Face Hub. By default we suggest to use CodeLlama.
|
44 |
|
45 |
To activate it, create a `.env.local` configuration file:
|
46 |
|
@@ -50,13 +50,14 @@ LLM_ENGINE="INFERENCE_API"
|
|
50 |
HF_API_TOKEN="Your Hugging Face token"
|
51 |
|
52 |
# codellama/CodeLlama-7b-hf" is used by default, but you can change this
|
53 |
-
# note: You should use a model able to generate JSON responses
|
|
|
54 |
HF_INFERENCE_API_MODEL="codellama/CodeLlama-7b-hf"
|
55 |
```
|
56 |
|
57 |
### Option 2: Use an Inference Endpoint URL
|
58 |
|
59 |
-
If
|
60 |
|
61 |
```bash
|
62 |
LLM_ENGINE="INFERENCE_ENDPOINT"
|
|
|
40 |
|
41 |
### Option 1: Use an Inference API model
|
42 |
|
43 |
+
This is a new option added recently, where you can use one of the models from the Hugging Face Hub. By default we suggest to use CodeLlama 34b as it will provide better results than the 7b model.
|
44 |
|
45 |
To activate it, create a `.env.local` configuration file:
|
46 |
|
|
|
50 |
HF_API_TOKEN="Your Hugging Face token"
|
51 |
|
52 |
# codellama/CodeLlama-7b-hf" is used by default, but you can change this
|
53 |
+
# note: You should use a model able to generate JSON responses,
|
54 |
+
# so it is storngly suggested to use at least the 34b model
|
55 |
HF_INFERENCE_API_MODEL="codellama/CodeLlama-7b-hf"
|
56 |
```
|
57 |
|
58 |
### Option 2: Use an Inference Endpoint URL
|
59 |
|
60 |
+
If you would like to run the AI Comic Factory on a private LLM running on the Hugging Face Inference Endpoint service, create a `.env.local` configuration file:
|
61 |
|
62 |
```bash
|
63 |
LLM_ENGINE="INFERENCE_ENDPOINT"
|