Upload folder using huggingface_hub
Browse files
model.yml
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
name:
|
2 |
model: llama3.1:8B
|
3 |
version: 1
|
4 |
|
@@ -17,8 +17,10 @@ stream: true # true | false
|
|
17 |
# Engine / Model Settings
|
18 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
19 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
20 |
-
engine:
|
21 |
-
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n
|
|
|
|
|
22 |
# Prompt template: Can only be retrieved from instruct model
|
23 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
24 |
-
# - Requires jinja format parser
|
|
|
1 |
+
name: llama3.1:8b-gguf-q4-km
|
2 |
model: llama3.1:8B
|
3 |
version: 1
|
4 |
|
|
|
17 |
# Engine / Model Settings
|
18 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
19 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
20 |
+
engine: llama-cpp
|
21 |
+
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n\
|
22 |
+
{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\
|
23 |
+
\n"
|
24 |
# Prompt template: Can only be retrieved from instruct model
|
25 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
26 |
+
# - Requires jinja format parser
|