Text Generation
Transformers
PyTorch
mpt
Composer
MosaicML
llm-foundry
custom_code
text-generation-inference
sam-mosaic atrott commited on
Commit
9f90bad
1 Parent(s): de18ef9

Update README.md (#4)

Browse files

- Update README.md (d51fa781e2bfad9f2a28c0d2987c7dee76e97d77)


Co-authored-by: Alex Trott <atrott@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -145,6 +145,11 @@ For more details on the pretraining process, see [MPT-7B](https://huggingface.co
145
 
146
  The data was tokenized using the [EleutherAI/gpt-neox-20b](https://huggingface.co/EleutherAI/gpt-neox-20b) tokenizer.
147
 
 
 
 
 
 
148
  ## Limitations and Biases
149
 
150
  _The following language is modified from [EleutherAI's GPT-NeoX-20B](https://huggingface.co/EleutherAI/gpt-neox-20b)_
 
145
 
146
  The data was tokenized using the [EleutherAI/gpt-neox-20b](https://huggingface.co/EleutherAI/gpt-neox-20b) tokenizer.
147
 
148
+ ### Training Configuration
149
+
150
+ This model was trained on 8 A100-80GBs for about 2 days using the [MosaicML Platform](https://www.mosaicml.com/platform).
151
+ The model was trained with sharded data parallelism using [FSDP](https://pytorch.org/docs/stable/fsdp.html) and used the [LION](https://arxiv.org/abs/2302.06675) optimizer.
152
+
153
  ## Limitations and Biases
154
 
155
  _The following language is modified from [EleutherAI's GPT-NeoX-20B](https://huggingface.co/EleutherAI/gpt-neox-20b)_