Text Generation
Transformers
PyTorch
mpt
Composer
MosaicML
llm-foundry
StreamingDatasets
custom_code
text-generation-inference
7 papers
lukasmoeller commited on
Commit
3d6ba6c
1 Parent(s): 06c1397

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -11,9 +11,16 @@ datasets:
11
  - togethercomputer/RedPajama-Data-1T
12
  - bigcode/the-stack
13
  - allenai/s2orc
 
14
  inference: false
15
  ---
16
 
 
 
 
 
 
 
17
  # MPT-7B
18
 
19
  MPT-7B is a decoder-style transformer pretrained from scratch on 1T tokens of English text and code.
 
11
  - togethercomputer/RedPajama-Data-1T
12
  - bigcode/the-stack
13
  - allenai/s2orc
14
+ - lukasmoeller/sail_preprocessed
15
  inference: false
16
  ---
17
 
18
+ # MPT-7B SAIL
19
+
20
+ This is a fine-tuned variant of MPT-7B, trained on the SAIL dataset (https://arxiv.org/abs/2305.15225). The preprocessed version can be found here: https://huggingface.co/datasets/lukasmoeller/sail_preprocessed
21
+
22
+ I may have forgotten to add EOD tokens at the end of the target, might retrain if anyone is interested.
23
+
24
  # MPT-7B
25
 
26
  MPT-7B is a decoder-style transformer pretrained from scratch on 1T tokens of English text and code.