sethuiyer commited on
Commit
d80838d
1 Parent(s): 1b3cd67

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -24,7 +24,7 @@ This metaphorically represents the model's depth, fluidity, and adaptability in
24
 
25
  It also perfectly fits the approach taken here - Depth Upscaling, inspired by SOLAR 10.7B.
26
 
27
- ## Nous LLM Evaluation (Version 1 - with ChatML Prompt Template)
28
  | Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
29
  |---------------------------------------------------------------|------:|------:|---------:|-------:|------:|
30
  |[Chikuma_10.7B](https://huggingface.co/sethuiyer/Chikuma_10.7B)| 42.41| 73.41| 56.69| 43.5| 54|
@@ -32,7 +32,7 @@ It also perfectly fits the approach taken here - Depth Upscaling, inspired by SO
32
  More details can be found [here](https://gist.github.com/sethuiyer/08b4498ed13a6dead38ad3a6f12e349a)
33
 
34
 
35
- ### Recommended Prompt Template
36
 
37
  ```text
38
  <|im_start|>GPT4 Correct system
@@ -45,11 +45,12 @@ Always use <|end_of_turn|> when you want to end the answer.<|im_end|>
45
  {{Input}}
46
  <|im_end|>GPT4 Correct Assistant:
47
  ```
48
- ChatML format also works well.
 
49
 
50
  ## Tested to work well in :
51
- 1. [text-generation-webui](https://github.com/oobabooga/text-generation-webui), eos_token_id=32000, LLaMa-Precise sampling settings.
52
- 2. `transformers` text generation pipeline, temperature=4.0, top_k=50, top_p=0.01, eos_token_id=32000.
53
 
54
 
55
  ## 🧩 Configuration
@@ -83,6 +84,6 @@ Tell me what is a large language model in under 250 words.
83
 
84
  messages = [{"role":"system", "content": sys_message}, {"role": "user", "content": question}]
85
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
86
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=4.0, top_k=50, top_p=0.01, eos_token_id=32000)
87
  print(outputs[0]["generated_text"])
88
  ```
 
24
 
25
  It also perfectly fits the approach taken here - Depth Upscaling, inspired by SOLAR 10.7B.
26
 
27
+ ## Nous LLM Evaluation (with ChatML Prompt Template)
28
  | Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
29
  |---------------------------------------------------------------|------:|------:|---------:|-------:|------:|
30
  |[Chikuma_10.7B](https://huggingface.co/sethuiyer/Chikuma_10.7B)| 42.41| 73.41| 56.69| 43.5| 54|
 
32
  More details can be found [here](https://gist.github.com/sethuiyer/08b4498ed13a6dead38ad3a6f12e349a)
33
 
34
 
35
+ ### Recommended Prompt Template (Experimental)
36
 
37
  ```text
38
  <|im_start|>GPT4 Correct system
 
45
  {{Input}}
46
  <|im_end|>GPT4 Correct Assistant:
47
  ```
48
+
49
+ ChatML also works, but make sure to add the sentence "Always use <|end_of_turn|> when you want to end the answer" as the default eos token is <|end_of_turn|>.
50
 
51
  ## Tested to work well in :
52
+ 1. [text-generation-webui](https://github.com/oobabooga/text-generation-webui), LLaMa-Precise sampling settings.
53
+ 2. `transformers` text generation pipeline, temperature=4.0, top_k=50, top_p=0.01.
54
 
55
 
56
  ## 🧩 Configuration
 
84
 
85
  messages = [{"role":"system", "content": sys_message}, {"role": "user", "content": question}]
86
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
87
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=4.0, top_k=50, top_p=0.01)
88
  print(outputs[0]["generated_text"])
89
  ```