Update README.md
Browse files
README.md
CHANGED
@@ -24,7 +24,7 @@ This metaphorically represents the model's depth, fluidity, and adaptability in
|
|
24 |
|
25 |
It also perfectly fits the approach taken here - Depth Upscaling, inspired by SOLAR 10.7B.
|
26 |
|
27 |
-
## Nous LLM Evaluation (
|
28 |
| Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
|
29 |
|---------------------------------------------------------------|------:|------:|---------:|-------:|------:|
|
30 |
|[Chikuma_10.7B](https://huggingface.co/sethuiyer/Chikuma_10.7B)| 42.41| 73.41| 56.69| 43.5| 54|
|
@@ -32,7 +32,7 @@ It also perfectly fits the approach taken here - Depth Upscaling, inspired by SO
|
|
32 |
More details can be found [here](https://gist.github.com/sethuiyer/08b4498ed13a6dead38ad3a6f12e349a)
|
33 |
|
34 |
|
35 |
-
### Recommended Prompt Template
|
36 |
|
37 |
```text
|
38 |
<|im_start|>GPT4 Correct system
|
@@ -45,11 +45,12 @@ Always use <|end_of_turn|> when you want to end the answer.<|im_end|>
|
|
45 |
{{Input}}
|
46 |
<|im_end|>GPT4 Correct Assistant:
|
47 |
```
|
48 |
-
|
|
|
49 |
|
50 |
## Tested to work well in :
|
51 |
-
1. [text-generation-webui](https://github.com/oobabooga/text-generation-webui),
|
52 |
-
2. `transformers` text generation pipeline, temperature=4.0, top_k=50, top_p=0.01
|
53 |
|
54 |
|
55 |
## 🧩 Configuration
|
@@ -83,6 +84,6 @@ Tell me what is a large language model in under 250 words.
|
|
83 |
|
84 |
messages = [{"role":"system", "content": sys_message}, {"role": "user", "content": question}]
|
85 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
86 |
-
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=4.0, top_k=50, top_p=0.01
|
87 |
print(outputs[0]["generated_text"])
|
88 |
```
|
|
|
24 |
|
25 |
It also perfectly fits the approach taken here - Depth Upscaling, inspired by SOLAR 10.7B.
|
26 |
|
27 |
+
## Nous LLM Evaluation (with ChatML Prompt Template)
|
28 |
| Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
|
29 |
|---------------------------------------------------------------|------:|------:|---------:|-------:|------:|
|
30 |
|[Chikuma_10.7B](https://huggingface.co/sethuiyer/Chikuma_10.7B)| 42.41| 73.41| 56.69| 43.5| 54|
|
|
|
32 |
More details can be found [here](https://gist.github.com/sethuiyer/08b4498ed13a6dead38ad3a6f12e349a)
|
33 |
|
34 |
|
35 |
+
### Recommended Prompt Template (Experimental)
|
36 |
|
37 |
```text
|
38 |
<|im_start|>GPT4 Correct system
|
|
|
45 |
{{Input}}
|
46 |
<|im_end|>GPT4 Correct Assistant:
|
47 |
```
|
48 |
+
|
49 |
+
ChatML also works, but make sure to add the sentence "Always use <|end_of_turn|> when you want to end the answer" as the default eos token is <|end_of_turn|>.
|
50 |
|
51 |
## Tested to work well in :
|
52 |
+
1. [text-generation-webui](https://github.com/oobabooga/text-generation-webui), LLaMa-Precise sampling settings.
|
53 |
+
2. `transformers` text generation pipeline, temperature=4.0, top_k=50, top_p=0.01.
|
54 |
|
55 |
|
56 |
## 🧩 Configuration
|
|
|
84 |
|
85 |
messages = [{"role":"system", "content": sys_message}, {"role": "user", "content": question}]
|
86 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
87 |
+
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=4.0, top_k=50, top_p=0.01)
|
88 |
print(outputs[0]["generated_text"])
|
89 |
```
|