|
|
|
id: smollm2 |
|
model: smollm2 |
|
name: smollm2 |
|
version: 1 |
|
|
|
|
|
|
|
|
|
stop: |
|
- <|im_end|> |
|
|
|
|
|
|
|
stream: true |
|
top_p: 0.9 |
|
temperature: 0.7 |
|
frequency_penalty: 0 |
|
presence_penalty: 0 |
|
max_tokens: 4096 |
|
seed: -1 |
|
dynatemp_range: 0 |
|
dynatemp_exponent: 1 |
|
top_k: 40 |
|
min_p: 0.05 |
|
tfs_z: 1 |
|
typ_p: 1 |
|
repeat_last_n: 64 |
|
repeat_penalty: 1 |
|
mirostat: false |
|
mirostat_tau: 5 |
|
mirostat_eta: 0.100000001 |
|
penalize_nl: false |
|
ignore_eos: false |
|
n_probs: 0 |
|
min_keep: 0 |
|
|
|
|
|
|
|
|
|
|
|
engine: llama-cpp |
|
ctx_len: 4096 |
|
ngl: 34 |
|
prompt_template: "<|im_start|>system |
|
{system_message}<|im_end|> |
|
<|im_start|>user |
|
{prompt}<|im_end|> |
|
<|im_start|>assistant" |
|
|
|
|
|
|