TheBloke commited on
Commit
0c23a00
1 Parent(s): 2ceade0

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -18
README.md CHANGED
@@ -5,16 +5,9 @@ license: llama2
5
  model_creator: Xwin-LM
6
  model_name: Xwin-LM 70B V0.1
7
  model_type: llama
8
- prompt_template: 'Below is an instruction that describes a task. Write a response
9
- that appropriately completes the request.
10
-
11
-
12
- ### Instruction:
13
-
14
- {prompt}
15
-
16
-
17
- ### Response:
18
 
19
  '
20
  quantized_by: TheBloke
@@ -75,15 +68,10 @@ Here is an incomplate list of clients and libraries that are known to support GG
75
  <!-- repositories-available end -->
76
 
77
  <!-- prompt-template start -->
78
- ## Prompt template: Alpaca
79
 
80
  ```
81
- Below is an instruction that describes a task. Write a response that appropriately completes the request.
82
-
83
- ### Instruction:
84
- {prompt}
85
-
86
- ### Response:
87
 
88
  ```
89
 
@@ -231,7 +219,7 @@ Windows Command Line users: You can set the environment variable by running `set
231
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
232
 
233
  ```shell
234
- ./main -ngl 32 -m xwin-lm-70b-v0.1.Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{prompt}\n\n### Response:"
235
  ```
236
 
237
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
 
5
  model_creator: Xwin-LM
6
  model_name: Xwin-LM 70B V0.1
7
  model_type: llama
8
+ prompt_template: 'A chat between a curious user and an artificial intelligence assistant.
9
+ The assistant gives helpful, detailed, and polite answers to the user''s questions.
10
+ USER: {prompt} ASSISTANT:
 
 
 
 
 
 
 
11
 
12
  '
13
  quantized_by: TheBloke
 
68
  <!-- repositories-available end -->
69
 
70
  <!-- prompt-template start -->
71
+ ## Prompt template: Vicuna
72
 
73
  ```
74
+ A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {prompt} ASSISTANT:
 
 
 
 
 
75
 
76
  ```
77
 
 
219
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
220
 
221
  ```shell
222
+ ./main -ngl 32 -m xwin-lm-70b-v0.1.Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {prompt} ASSISTANT:"
223
  ```
224
 
225
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.