Update README.md
#76
by
ybelkada
- opened
README.md
CHANGED
@@ -2444,11 +2444,8 @@ Initial prompting experiments using interim checkpoints: https://huggingface.co/
|
|
2444 |
|
2445 |
|
2446 |
## Original checkpoints
|
2447 |
-
|
2448 |
-
The checkpoints in this repo correspond to the HuggingFace Transformers format. If you want to use our fork of [Megatron-DeepSpeed](https://github.com/bigscience-workshop/Megatron-DeepSpeed) that the model was trained with, you'd want to use [this repo instead](https://huggingface.co/bigscience/bloom-optimizer-states).
|
2449 |
-
|
2450 |
---
|
2451 |
-
|
2452 |
# Model Card Authors
|
2453 |
*Ordered roughly chronologically and by amount of time spent.*
|
2454 |
|
|
|
2444 |
|
2445 |
|
2446 |
## Original checkpoints
|
2447 |
+
The checkpoints in this repo correspond to the HuggingFace Transformers format. If you want to use our fork of [Megatron-DeepSpeed](https://github.com/bigscience-workshop/Megatron-DeepSpeed) that the model was trained with, you’d want to use [this repo instead](https://huggingface.co/bigscience/bloom-megatron-deepspeed) that contains full Megatron-DeepSpeed checkpoints and optimizer states.
|
|
|
|
|
2448 |
---
|
|
|
2449 |
# Model Card Authors
|
2450 |
*Ordered roughly chronologically and by amount of time spent.*
|
2451 |
|