Text Generation
Transformers
English
TheBloke commited on
Commit
79634b4
1 Parent(s): 98a7646

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -2
README.md CHANGED
@@ -1,6 +1,12 @@
1
  ---
2
  inference: false
3
- license: other
 
 
 
 
 
 
4
  ---
5
 
6
  <!-- header start -->
@@ -34,6 +40,19 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
34
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/orca_mini_v2_7B-GGML)
35
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/psmathur/orca_mini_v2_7b)
36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  <!-- compatibility_ggml start -->
38
  ## Compatibility
39
 
@@ -87,7 +106,7 @@ Refer to the Provided Files table below to see what files use which methods, and
87
  I use the following command line; adjust for your tastes and needs:
88
 
89
  ```
90
- ./main -t 10 -ngl 32 -m orca-mini-v2_7b.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "### Instruction: Write a story about llamas\n### Response:"
91
  ```
92
  If you're able to use full GPU offloading, you should use `-t 1` to get best performance.
93
 
 
1
  ---
2
  inference: false
3
+ license: cc-by-nc-sa-4.0
4
+ language:
5
+ - en
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
+ datasets:
9
+ - psmathur/orca_minis_uncensored_dataset
10
  ---
11
 
12
  <!-- header start -->
 
40
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/orca_mini_v2_7B-GGML)
41
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/psmathur/orca_mini_v2_7b)
42
 
43
+ ## Prompt template
44
+ ```
45
+ ### System:
46
+ You are an AI assistant that follows instruction extremely well. Help as much as you can.
47
+
48
+ ### User:
49
+ prompt
50
+
51
+ ### Input:
52
+
53
+ ### Response:
54
+ ```
55
+
56
  <!-- compatibility_ggml start -->
57
  ## Compatibility
58
 
 
106
  I use the following command line; adjust for your tastes and needs:
107
 
108
  ```
109
+ ./main -t 10 -ngl 32 -m orca-mini-v2_7b.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "### System: You are an AI assistant that follows instruction extremely well. Help as much as you can.### User: Write a story about llamas\n### Response:"
110
  ```
111
  If you're able to use full GPU offloading, you should use `-t 1` to get best performance.
112