TheBloke commited on
Commit
04f23f8
1 Parent(s): 66757f6

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -11,7 +11,7 @@ model_creator: Open-Orca
11
  model_name: OpenOrca x OpenChat - Preview2 - 13B
12
  model_type: llama
13
  pipeline_tag: text-generation
14
- prompt_template: 'user: {prompt}<|end_of_turn|>assistant:
15
 
16
  '
17
  quantized_by: TheBloke
@@ -56,10 +56,10 @@ Multiple GPTQ parameter permutations are provided; see Provided Files below for
56
  <!-- repositories-available end -->
57
 
58
  <!-- prompt-template start -->
59
- ## Prompt template: openchat llama2 v1
60
 
61
  ```
62
- user: {prompt}<|end_of_turn|>assistant:
63
 
64
  ```
65
 
@@ -175,7 +175,7 @@ model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
175
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
176
 
177
  prompt = "Tell me about AI"
178
- prompt_template=f'''user: {prompt}<|end_of_turn|>assistant:
179
 
180
  '''
181
 
 
11
  model_name: OpenOrca x OpenChat - Preview2 - 13B
12
  model_type: llama
13
  pipeline_tag: text-generation
14
+ prompt_template: 'GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
15
 
16
  '
17
  quantized_by: TheBloke
 
56
  <!-- repositories-available end -->
57
 
58
  <!-- prompt-template start -->
59
+ ## Prompt template: OpenChat
60
 
61
  ```
62
+ GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
63
 
64
  ```
65
 
 
175
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
176
 
177
  prompt = "Tell me about AI"
178
+ prompt_template=f'''GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
179
 
180
  '''
181