Locutusque commited on
Commit
fc9b621
1 Parent(s): 78b518e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -2,6 +2,7 @@
2
  license: apache-2.0
3
  datasets:
4
  - Skylion007/openwebtext
 
5
  language:
6
  - en
7
  pipeline_tag: text-generation
@@ -14,7 +15,7 @@ inference:
14
  max_new_tokens: 250
15
  repetition_penalty: 1.176
16
  ---
17
- A pre-trained language model, based on the Mistral 7B model, has been scaled down to approximately 248 million parameters. This model has been trained on 7,320,000 examples. This model isn't intended for direct use but for fine-tuning on a downstream task.
18
  This model should have a context length of around 32,768 tokens. Safe serialization has been removed due to issues saving model weights.
19
 
20
  During evaluation on InstructMix, this model achieved an average perplexity score of 6.3. More epochs are planned for this model on different datasets.
 
2
  license: apache-2.0
3
  datasets:
4
  - Skylion007/openwebtext
5
+ - JeanKaddour/minipile
6
  language:
7
  - en
8
  pipeline_tag: text-generation
 
15
  max_new_tokens: 250
16
  repetition_penalty: 1.176
17
  ---
18
+ A pre-trained language model, based on the Mistral 7B model, has been scaled down to approximately 248 million parameters. This model has been trained on 7,488,000 examples. This model isn't intended for direct use but for fine-tuning on a downstream task.
19
  This model should have a context length of around 32,768 tokens. Safe serialization has been removed due to issues saving model weights.
20
 
21
  During evaluation on InstructMix, this model achieved an average perplexity score of 6.3. More epochs are planned for this model on different datasets.