flyingfishinwater commited on
Commit
9e2d5a6
·
verified ·
1 Parent(s): 8da8581

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -35
README.md CHANGED
@@ -1,8 +1,46 @@
1
  ---
2
  license: apache-2.0
3
  ---
4
- # LiteLlama
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens. It's best for testing.
7
 
8
  **Model Intention:** This is a 460 parameters' very small model for test purpose only
@@ -22,11 +60,10 @@ It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens.
22
  **Context Length:** 1024 tokens
23
 
24
  **Prompt Format:**
25
-
26
  ```
27
  <human>: {{prompt}}
28
  <bot>:
29
- ```
30
 
31
  **Template Name:** TinyLlama
32
 
@@ -40,7 +77,6 @@ It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens.
40
  ---
41
 
42
  # TinyLlama-1.1B-chat
43
-
44
  The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, we can achieve this within a span of just 90 days using 16 A100-40G GPUs. The training has started on 2023-09-01.
45
 
46
  **Model Intention:** It's good for question & answer.
@@ -60,10 +96,9 @@ The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens.
60
  **Context Length:** 4096 tokens
61
 
62
  **Prompt Format:**
63
-
64
  ```
65
  <|system|>You are a friendly chatbot who always responds in the style of a pirate.</s><|user|>{{prompt}}</s><|assistant|>
66
- ```
67
 
68
  **Template Name:** TinyLlama
69
 
@@ -77,7 +112,6 @@ The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens.
77
  ---
78
 
79
  # Mistral 7B v0.2
80
-
81
  The Mistral-7B-v0.2 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.2 outperforms Llama 2 13B on all benchmarks we tested.
82
 
83
  **Model Intention:** It's a 7B large model for Q&A purpose. But it requires a high-end device to run.
@@ -97,10 +131,9 @@ The Mistral-7B-v0.2 Large Language Model (LLM) is a pretrained generative text m
97
  **Context Length:** 4096 tokens
98
 
99
  **Prompt Format:**
100
-
101
  ```
102
  <s>[INST]{{prompt}}[/INST]</s>
103
- ```
104
 
105
  **Template Name:** Mistral
106
 
@@ -114,7 +147,6 @@ The Mistral-7B-v0.2 Large Language Model (LLM) is a pretrained generative text m
114
  ---
115
 
116
  # OpenChat 3.5
117
-
118
  OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning. Our models learn from mixed-quality data without preference labels, delivering exceptional performance on par with ChatGPT, even with a 7B model. Despite our simple approach, we are committed to developing a high-performance, commercially viable, open-source large language model, and we continue to make significant strides toward this vision.
119
 
120
  **Model Intention:** It's a 7B large model and performs really good for Q&A. But it requires a high-end device to run.
@@ -134,10 +166,9 @@ OpenChat is an innovative library of open-source language models, fine-tuned wit
134
  **Context Length:** 4096 tokens
135
 
136
  **Prompt Format:**
137
-
138
  ```
139
  <s>[INST]{{prompt}}[/INST]</s>
140
- ```
141
 
142
  **Template Name:** Mistral
143
 
@@ -151,7 +182,6 @@ OpenChat is an innovative library of open-source language models, fine-tuned wit
151
  ---
152
 
153
  # Phi-2
154
-
155
  Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the same data sources as Phi-1.5, augmented with a new data source that consists of various NLP synthetic texts and filtered websites (for safety and educational value). When assessed against benchmarks testing common sense, language understanding, and logical reasoning, Phi-2 showcased a nearly state-of-the-art performance among models with less than 13 billion parameters.
156
 
157
  **Model Intention:** It's a 2.7B model and is intended for QA, chat, and code purposes
@@ -171,11 +201,10 @@ Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the sam
171
  **Context Length:** 4096 tokens
172
 
173
  **Prompt Format:**
174
-
175
  ```
176
  Instruct: {{prompt}}
177
- Output:
178
- ```
179
 
180
  **Template Name:** PHI
181
 
@@ -189,7 +218,6 @@ Output:
189
  ---
190
 
191
  # Yi 6B Chat
192
-
193
  The Yi series models are the next generation of open-source large language models trained from scratch by 01.AI. Targeted as a bilingual language model and trained on 3T multilingual corpus, the Yi series models become one of the strongest LLM worldwide, showing promise in language understanding, commonsense reasoning, reading comprehension, and more. For example, For English language capability, the Yi series models ranked 2nd (just behind GPT-4), outperforming other LLMs (such as LLaMA2-chat-70B, Claude 2, and ChatGPT) on the AlpacaEval Leaderboard in Dec 2023. For Chinese language capability, the Yi series models landed in 2nd place (following GPT-4), surpassing other LLMs (such as Baidu ERNIE, Qwen, and Baichuan) on the SuperCLUE in Oct 2023.
194
 
195
  **Model Intention:** It's a 6B model and can understand English and Chinese. It's good for QA and Chat
@@ -209,14 +237,13 @@ The Yi series models are the next generation of open-source large language model
209
  **Context Length:** 200000 tokens
210
 
211
  **Prompt Format:**
212
-
213
  ```
214
  <|im_start|>user
215
  <|im_end|>
216
  {{prompt}}
217
  <|im_start|>assistant
218
 
219
- ```
220
 
221
  **Template Name:** yi
222
 
@@ -230,7 +257,6 @@ The Yi series models are the next generation of open-source large language model
230
  ---
231
 
232
  # Google Gemma 2B
233
-
234
  Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Developed by Google DeepMind and other teams across Google, Gemma is named after the Latin gemma, meaning 'precious stone.' The Gemma model weights are supported by developer tools that promote innovation, collaboration, and the responsible use of artificial intelligence (AI).
235
 
236
  **Model Intention:** It's a 2B large model for Q&A purpose. But it requires a high-end device to run.
@@ -250,13 +276,12 @@ Gemma is a family of lightweight, state-of-the-art open models built from the sa
250
  **Context Length:** 8192 tokens
251
 
252
  **Prompt Format:**
253
-
254
  ```
255
  <bos><start_of_turn>user
256
  {{prompt}}<end_of_turn>
257
  <start_of_turn>model
258
 
259
- ```
260
 
261
  **Template Name:** gemma
262
 
@@ -270,7 +295,6 @@ Gemma is a family of lightweight, state-of-the-art open models built from the sa
270
  ---
271
 
272
  # StarCoder2 3B
273
-
274
  StarCoder2-3B model is a 3B parameter model trained on 17 programming languages from The Stack v2, with opt-out requests excluded. The model uses Grouped Query Attention, a context window of 16,384 tokens with a sliding window attention of 4,096 tokens, and was trained using the Fill-in-the-Middle objective on 3+ trillion tokens
275
 
276
  **Model Intention:** The model is good at 17 programming languages. It can help you resolve programming requirements
@@ -290,12 +314,11 @@ StarCoder2-3B model is a 3B parameter model trained on 17 programming languages
290
  **Context Length:** 8192 tokens
291
 
292
  **Prompt Format:**
293
-
294
  ```
295
  ### Instruction
296
  {{prompt}}### Response
297
 
298
- ```
299
 
300
  **Template Name:** starcoder
301
 
@@ -309,7 +332,6 @@ StarCoder2-3B model is a 3B parameter model trained on 17 programming languages
309
  ---
310
 
311
  # Chinese Tiny LLM 2B
312
-
313
  Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在中文语料库上进行预训练和微调,提供了对潜在偏见、中文语言能力和多语言适应性的重要洞见。
314
 
315
  **Model Intention:** 这是一个参数规模2B的中文模型,具有很好的中文理解和应答能力
@@ -329,14 +351,13 @@ Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在
329
  **Context Length:** 4096 tokens
330
 
331
  **Prompt Format:**
332
-
333
  ```
334
  <|im_start|>user
335
  {{prompt}}
336
  <|im_end|>
337
  <|im_start|>assistant
338
 
339
- ```
340
 
341
  **Template Name:** chatml
342
 
@@ -350,7 +371,6 @@ Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在
350
  ---
351
 
352
  # Dophin 2.8 Mistralv02 7B
353
-
354
  This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensored model and supports a variety of instruction, conversational, and coding skills.
355
 
356
  **Model Intention:** It's a uncensored and good skilled English modal best for high performance iPhone, iPad & Mac
@@ -370,14 +390,13 @@ This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensor
370
  **Context Length:** 16384 tokens
371
 
372
  **Prompt Format:**
373
-
374
  ```
375
  <|im_start|>user
376
  {{prompt}}
377
  <|im_end|>
378
  <|im_start|>assistant
379
 
380
- ```
381
 
382
  **Template Name:** chatml
383
 
@@ -391,7 +410,6 @@ This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensor
391
  ---
392
 
393
  # WizardLM-2 7B
394
-
395
  The WizardLM-2 is one of the next generation state-of-the-art large language models, which have improved performance on complex chat, multilingual, reasoning and agent.
396
 
397
  **Model Intention:** It's a state-of-the-art large language model with improved performance on complex chat, multilingual, reasoning and agent.
@@ -411,12 +429,11 @@ The WizardLM-2 is one of the next generation state-of-the-art large language mod
411
  **Context Length:** 32768 tokens
412
 
413
  **Prompt Format:**
414
-
415
  ```
416
  A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
417
  USER: {{prompt}}
418
  ASSISTANT:
419
- ```
420
 
421
  **Template Name:** chatml
422
 
@@ -424,4 +441,7 @@ ASSISTANT:
424
 
425
  **Add EOS Token:** No
426
 
427
- **Parse Special Tokens:** Yes
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ # Llama3 8B
5
+ Llama 3 is the latest and most advanced LLM trained over 15T tokens, which improves its comprehension and handling of complex language nuances. It features an extended context window of 8k tokens allowing the model to access more information from lengthy passages for more informed decision-making.
6
+
7
+ **Model Intention:** The latest Llama 3 enabling more accurate and informative responses to complex queries in both English and multilingual contexts.
8
+
9
+ **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf?download=true)
10
+
11
+ **Model Info URL:** [https://huggingface.co/meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
12
+
13
+ **Model License:** [License Info](https://llama.meta.com/llama3/license/)
14
+
15
+ **Model Description:** Llama 3 is the latest and most advanced LLM trained over 15T tokens, which improves its comprehension and handling of complex language nuances. It features an extended context window of 8k tokens allowing the model to access more information from lengthy passages for more informed decision-making.
16
+
17
+ **Developer:** [https://llama.meta.com/](https://llama.meta.com/)
18
+
19
+ **File Size:** 4921 MB
20
+
21
+ **Context Length:** 8192 tokens
22
+
23
+ **Prompt Format:**
24
+ ```
25
+ <|begin_of_text|><|start_header_id|>user<|end_header_id|>
26
+
27
+ {{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
28
+
29
+
30
+ ```
31
 
32
+ **Template Name:** llama
33
+
34
+ **Add BOS Token:** Yes
35
+
36
+ **Add EOS Token:** No
37
+
38
+ **Parse Special Tokens:** Yes
39
+
40
+
41
+ ---
42
+
43
+ # LiteLlama
44
  It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens. It's best for testing.
45
 
46
  **Model Intention:** This is a 460 parameters' very small model for test purpose only
 
60
  **Context Length:** 1024 tokens
61
 
62
  **Prompt Format:**
 
63
  ```
64
  <human>: {{prompt}}
65
  <bot>:
66
+ ```
67
 
68
  **Template Name:** TinyLlama
69
 
 
77
  ---
78
 
79
  # TinyLlama-1.1B-chat
 
80
  The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, we can achieve this within a span of just 90 days using 16 A100-40G GPUs. The training has started on 2023-09-01.
81
 
82
  **Model Intention:** It's good for question & answer.
 
96
  **Context Length:** 4096 tokens
97
 
98
  **Prompt Format:**
 
99
  ```
100
  <|system|>You are a friendly chatbot who always responds in the style of a pirate.</s><|user|>{{prompt}}</s><|assistant|>
101
+ ```
102
 
103
  **Template Name:** TinyLlama
104
 
 
112
  ---
113
 
114
  # Mistral 7B v0.2
 
115
  The Mistral-7B-v0.2 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.2 outperforms Llama 2 13B on all benchmarks we tested.
116
 
117
  **Model Intention:** It's a 7B large model for Q&A purpose. But it requires a high-end device to run.
 
131
  **Context Length:** 4096 tokens
132
 
133
  **Prompt Format:**
 
134
  ```
135
  <s>[INST]{{prompt}}[/INST]</s>
136
+ ```
137
 
138
  **Template Name:** Mistral
139
 
 
147
  ---
148
 
149
  # OpenChat 3.5
 
150
  OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning. Our models learn from mixed-quality data without preference labels, delivering exceptional performance on par with ChatGPT, even with a 7B model. Despite our simple approach, we are committed to developing a high-performance, commercially viable, open-source large language model, and we continue to make significant strides toward this vision.
151
 
152
  **Model Intention:** It's a 7B large model and performs really good for Q&A. But it requires a high-end device to run.
 
166
  **Context Length:** 4096 tokens
167
 
168
  **Prompt Format:**
 
169
  ```
170
  <s>[INST]{{prompt}}[/INST]</s>
171
+ ```
172
 
173
  **Template Name:** Mistral
174
 
 
182
  ---
183
 
184
  # Phi-2
 
185
  Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the same data sources as Phi-1.5, augmented with a new data source that consists of various NLP synthetic texts and filtered websites (for safety and educational value). When assessed against benchmarks testing common sense, language understanding, and logical reasoning, Phi-2 showcased a nearly state-of-the-art performance among models with less than 13 billion parameters.
186
 
187
  **Model Intention:** It's a 2.7B model and is intended for QA, chat, and code purposes
 
201
  **Context Length:** 4096 tokens
202
 
203
  **Prompt Format:**
 
204
  ```
205
  Instruct: {{prompt}}
206
+ Output:
207
+ ```
208
 
209
  **Template Name:** PHI
210
 
 
218
  ---
219
 
220
  # Yi 6B Chat
 
221
  The Yi series models are the next generation of open-source large language models trained from scratch by 01.AI. Targeted as a bilingual language model and trained on 3T multilingual corpus, the Yi series models become one of the strongest LLM worldwide, showing promise in language understanding, commonsense reasoning, reading comprehension, and more. For example, For English language capability, the Yi series models ranked 2nd (just behind GPT-4), outperforming other LLMs (such as LLaMA2-chat-70B, Claude 2, and ChatGPT) on the AlpacaEval Leaderboard in Dec 2023. For Chinese language capability, the Yi series models landed in 2nd place (following GPT-4), surpassing other LLMs (such as Baidu ERNIE, Qwen, and Baichuan) on the SuperCLUE in Oct 2023.
222
 
223
  **Model Intention:** It's a 6B model and can understand English and Chinese. It's good for QA and Chat
 
237
  **Context Length:** 200000 tokens
238
 
239
  **Prompt Format:**
 
240
  ```
241
  <|im_start|>user
242
  <|im_end|>
243
  {{prompt}}
244
  <|im_start|>assistant
245
 
246
+ ```
247
 
248
  **Template Name:** yi
249
 
 
257
  ---
258
 
259
  # Google Gemma 2B
 
260
  Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Developed by Google DeepMind and other teams across Google, Gemma is named after the Latin gemma, meaning 'precious stone.' The Gemma model weights are supported by developer tools that promote innovation, collaboration, and the responsible use of artificial intelligence (AI).
261
 
262
  **Model Intention:** It's a 2B large model for Q&A purpose. But it requires a high-end device to run.
 
276
  **Context Length:** 8192 tokens
277
 
278
  **Prompt Format:**
 
279
  ```
280
  <bos><start_of_turn>user
281
  {{prompt}}<end_of_turn>
282
  <start_of_turn>model
283
 
284
+ ```
285
 
286
  **Template Name:** gemma
287
 
 
295
  ---
296
 
297
  # StarCoder2 3B
 
298
  StarCoder2-3B model is a 3B parameter model trained on 17 programming languages from The Stack v2, with opt-out requests excluded. The model uses Grouped Query Attention, a context window of 16,384 tokens with a sliding window attention of 4,096 tokens, and was trained using the Fill-in-the-Middle objective on 3+ trillion tokens
299
 
300
  **Model Intention:** The model is good at 17 programming languages. It can help you resolve programming requirements
 
314
  **Context Length:** 8192 tokens
315
 
316
  **Prompt Format:**
 
317
  ```
318
  ### Instruction
319
  {{prompt}}### Response
320
 
321
+ ```
322
 
323
  **Template Name:** starcoder
324
 
 
332
  ---
333
 
334
  # Chinese Tiny LLM 2B
 
335
  Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在中文语料库上进行预训练和微调,提供了对潜在偏见、中文语言能力和多语言适应性的重要洞见。
336
 
337
  **Model Intention:** 这是一个参数规模2B的中文模型,具有很好的中文理解和应答能力
 
351
  **Context Length:** 4096 tokens
352
 
353
  **Prompt Format:**
 
354
  ```
355
  <|im_start|>user
356
  {{prompt}}
357
  <|im_end|>
358
  <|im_start|>assistant
359
 
360
+ ```
361
 
362
  **Template Name:** chatml
363
 
 
371
  ---
372
 
373
  # Dophin 2.8 Mistralv02 7B
 
374
  This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensored model and supports a variety of instruction, conversational, and coding skills.
375
 
376
  **Model Intention:** It's a uncensored and good skilled English modal best for high performance iPhone, iPad & Mac
 
390
  **Context Length:** 16384 tokens
391
 
392
  **Prompt Format:**
 
393
  ```
394
  <|im_start|>user
395
  {{prompt}}
396
  <|im_end|>
397
  <|im_start|>assistant
398
 
399
+ ```
400
 
401
  **Template Name:** chatml
402
 
 
410
  ---
411
 
412
  # WizardLM-2 7B
 
413
  The WizardLM-2 is one of the next generation state-of-the-art large language models, which have improved performance on complex chat, multilingual, reasoning and agent.
414
 
415
  **Model Intention:** It's a state-of-the-art large language model with improved performance on complex chat, multilingual, reasoning and agent.
 
429
  **Context Length:** 32768 tokens
430
 
431
  **Prompt Format:**
 
432
  ```
433
  A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
434
  USER: {{prompt}}
435
  ASSISTANT:
436
+ ```
437
 
438
  **Template Name:** chatml
439
 
 
441
 
442
  **Add EOS Token:** No
443
 
444
+ **Parse Special Tokens:** Yes
445
+
446
+
447
+ ---