jacobfulano commited on
Commit
b1188a4
1 Parent(s): 02a80c2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -1
README.md CHANGED
@@ -85,7 +85,13 @@ model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b', con
85
  model.to(device='cuda:0')
86
  ```
87
 
88
- The model size is approximately 13 GB total in two shards.
 
 
 
 
 
 
89
 
90
  This model was trained with the [EleutherAI/gpt-neox-20b](https://huggingface.co/EleutherAI/gpt-neox-20b) tokenizer.
91
 
 
85
  model.to(device='cuda:0')
86
  ```
87
 
88
+ Although the model was trained with a sequence length of 2048, ALiBi enables users to increase the maximum sequence length during finetuning and/or deployment. For example:
89
+
90
+ ```python
91
+ config = transformers.AutoConfig.from_pretrained('mosaicml/mpt-7b', trust_remote_code=True)
92
+ config.update({"max_seq_len": 4096})
93
+ model = transformers.AutoModelForCausalLM.from_pretrained('mosaicml/mpt-7b', config=config, trust_remote_code=True)
94
+ ```
95
 
96
  This model was trained with the [EleutherAI/gpt-neox-20b](https://huggingface.co/EleutherAI/gpt-neox-20b) tokenizer.
97