Transformers
fastai
English
Inference Endpoints
File size: 9,091 Bytes
6540589
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
[
  {
    "order": "a",
    "md5sum": "e8d47924f433bd561cb5244557147793",
    "name": "Wizard v1.1",
    "filename": "wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin",
    "filesize": "7323310848",
    "ramrequired": "16",
    "parameters": "13 billion",
    "quant": "q4_0",
    "type": "LLaMA",
    "systemPrompt": " ",
    "description": "<strong>Best overall model</strong><br><ul><li>Instruction based<li>Gives very long responses<li>Finetuned with only 1k of high-quality data<li>Trained by Microsoft and Peking University<li>Cannot be used commercially</ul"
  },
  {
    "order": "b",
    "md5sum": "725f148218a65ce8ebcc724e52f31b49",
    "name": "GPT4All Falcon",
    "filename": "ggml-model-gpt4all-falcon-q4_0.bin",
    "filesize": "4061641216",
    "requires": "2.4.9",
    "ramrequired": "8",
    "parameters": "7 billion",
    "quant": "q4_0",
    "type": "Falcon",
    "systemPrompt": " ",
    "description": "<strong>Best overall smaller model</strong><br><ul><li>Fast responses</li><li>Instruction based</li><li>Trained by TII<li>Finetuned by Nomic AI<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/nomic-ai/gpt4all-falcon-ggml/resolve/main/ggml-model-gpt4all-falcon-q4_0.bin",
    "promptTemplate": "### Instruction:\n%1\n### Response:\n"
  },
  {
    "order": "c",
    "md5sum": "4acc146dd43eb02845c233c29289c7c5",
    "name": "Hermes",
    "filename": "nous-hermes-13b.ggmlv3.q4_0.bin",
    "filesize": "8136777088",
    "requires": "2.4.7",
    "ramrequired": "16",
    "parameters": "13 billion",
    "quant": "q4_0",
    "type": "LLaMA",
    "systemPrompt": " ",
    "description": "<strong>Extremely good model</strong><br><ul><li>Instruction based<li>Gives long responses<li>Curated with 300,000 uncensored instructions<li>Trained by Nous Research<li>Cannot be used commercially</ul>",
    "url": "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML/resolve/main/nous-hermes-13b.ggmlv3.q4_0.bin",
    "promptTemplate": "### Instruction:\n%1\n### Response:\n"
  },
  {
    "order": "f",
    "md5sum": "11d9f060ca24575a2c303bdc39952486",
    "name": "Snoozy",
    "filename": "GPT4All-13B-snoozy.ggmlv3.q4_0.bin",
    "filesize": "8136770688",
    "requires": "2.4.7",
    "ramrequired": "16",
    "parameters": "13 billion",
    "quant": "q4_0",
    "type": "LLaMA",
    "systemPrompt": " ",
    "description": "<strong>Very good overall model</strong><br><ul><li>Instruction based<li>Based on the same dataset as Groovy<li>Slower than Groovy, with higher quality responses<li>Trained by Nomic AI<li>Cannot be used commercially</ul>",
    "url": "https://huggingface.co/TheBloke/GPT4All-13B-snoozy-GGML/resolve/main/GPT4All-13B-snoozy.ggmlv3.q4_0.bin"
  },
  {
    "order": "h",
    "md5sum": "e64e74375ce9d36a3d0af3db1523fd0a",
    "name": "Mini Orca",
    "filename": "orca-mini-7b.ggmlv3.q4_0.bin",
    "filesize": "3791749248",
    "requires": "2.4.7",
    "ramrequired": "8",
    "parameters": "7 billion",
    "quant": "q4_0",
    "type": "OpenLLaMa",
    "description": "<strong>New model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/TheBloke/orca_mini_7B-GGML/resolve/main/orca-mini-7b.ggmlv3.q4_0.bin",
    "promptTemplate": "### User:\n%1\n### Response:\n",
    "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
  },
  {
    "order": "i",
    "md5sum": "6a087f7f4598fad0bb70e6cb4023645e",
    "name": "Mini Orca (Small)",
    "filename": "orca-mini-3b.ggmlv3.q4_0.bin",
    "filesize": "1928446208",
    "requires": "2.4.7",
    "ramrequired": "4",
    "parameters": "3 billion",
    "quant": "q4_0",
    "type": "OpenLLaMa",
    "description": "<strong>Small version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/TheBloke/orca_mini_3B-GGML/resolve/main/orca-mini-3b.ggmlv3.q4_0.bin",
    "promptTemplate": "### User:\n%1\n### Response:\n",
    "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
  },
  {
    "order": "j",
    "md5sum": "959b7f65b2d12fd1e3ff99e7493c7a3a",
    "name": "Mini Orca (Large)",
    "filename": "orca-mini-13b.ggmlv3.q4_0.bin",
    "filesize": "7323329152",
    "requires": "2.4.7",
    "ramrequired": "16",
    "parameters": "13 billion",
    "quant": "q4_0",
    "type": "OpenLLaMa",
    "description": "<strong>Largest version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/TheBloke/orca_mini_13B-GGML/resolve/main/orca-mini-13b.ggmlv3.q4_0.bin",
    "promptTemplate": "### User:\n%1\n### Response:\n",
    "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
  },
  {
    "order": "r",
    "md5sum": "489d21fd48840dcb31e5f92f453f3a20",
    "name": "Wizard Uncensored",
    "filename": "wizardLM-13B-Uncensored.ggmlv3.q4_0.bin",
    "filesize": "8136777088",
    "requires": "2.4.7",
    "ramrequired": "16",
    "parameters": "13 billion",
    "quant": "q4_0",
    "type": "LLaMA",
    "systemPrompt": " ",
    "description": "<strong>Trained on uncensored assistant data and instruction data</strong><br><ul><li>Instruction based<li>Cannot be used commercially</ul>",
    "url": "https://huggingface.co/TheBloke/WizardLM-13B-Uncensored-GGML/resolve/main/wizardLM-13B-Uncensored.ggmlv3.q4_0.bin"
  },
  {
    "order": "s",
    "md5sum": "615890cb571fcaa0f70b2f8d15ef809e",
    "disableGUI": "true",
    "name": "Replit",
    "filename": "ggml-replit-code-v1-3b.bin",
    "filesize": "5202046853",
    "requires": "2.4.7",
    "ramrequired": "4",
    "parameters": "3 billion",
    "quant": "f16",
    "type": "Replit",
    "systemPrompt": " ",
    "promptTemplate": "%1",
    "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/nomic-ai/ggml-replit-code-v1-3b/resolve/main/ggml-replit-code-v1-3b.bin"
  },
  {
    "order": "t",
    "md5sum": "031bb5d5722c08d13e3e8eaf55c37391",
    "disableGUI": "true",
    "name": "Bert",
    "filename": "ggml-all-MiniLM-L6-v2-f16.bin",
    "filesize": "45521167",
    "requires": "2.4.14",
    "ramrequired": "1",
    "parameters": "1 million",
    "quant": "f16",
    "type": "Bert",
    "systemPrompt": " ",
    "description": "<strong>Sbert</strong><br><ul><li>For embeddings"
  },
  {
    "order": "u",
    "md5sum": "379ee1bab9a7a9c27c2314daa097528e",
    "disableGUI": "true",
    "name": "Starcoder (Small)",
    "filename": "starcoderbase-3b-ggml.bin",
    "filesize": "7503121552",
    "requires": "2.4.14",
    "ramrequired": "8",
    "parameters": "3 billion",
    "quant": "f16",
    "type": "Starcoder",
    "systemPrompt": " ",
    "promptTemplate": "%1",
    "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based</ul>"
  },
  {
    "order": "w",
    "md5sum": "f981ab8fbd1ebbe4932ddd667c108ba7",
    "disableGUI": "true",
    "name": "Starcoder",
    "filename": "starcoderbase-7b-ggml.bin",
    "filesize": "17860448016",
    "requires": "2.4.14",
    "ramrequired": "16",
    "parameters": "7 billion",
    "quant": "f16",
    "type": "Starcoder",
    "systemPrompt": " ",
    "promptTemplate": "%1",
    "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based</ul>"
  },
  {
    "order": "w",
    "md5sum": "c7ebc61eec1779bddae1f2bcbf2007cc",
    "name": "Llama-2-7B Chat",
    "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
    "filesize": "3791725184",
    "requires": "2.4.14",
    "ramrequired": "8",
    "parameters": "7 billion",
    "quant": "q4_0",
    "type": "LLaMA2",
    "description": "<strong>New LLaMA2 model from Meta AI.</strong><br><ul><li>Fine-tuned for dialogue.<li>static model trained on an offline dataset<li>RLHF dataset<li>Licensed for commercial use</ul>",
    "url": "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/resolve/main/llama-2-7b-chat.ggmlv3.q4_0.bin",
    "promptTemplate": "[INST] %1 [/INST] ",
    "systemPrompt": "[INST]<<SYS>>You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature. If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.<</SYS>>[/INST] "
  }
]