Update README.md
#3
by
ivas-tri
- opened
README.md
CHANGED
@@ -93,7 +93,7 @@ We follow their training recipe and release our version of Mamba-7B.
|
|
93 |
|
94 |
## Training Details
|
95 |
- Mamba-7B was trained using AWS SageMaker on 128 H100 80GB GPUs.
|
96 |
-
- Training began in March 2024 and lasted
|
97 |
| **Hyperparameter** | **Value** |
|
98 |
|--------------------|------------|
|
99 |
| Precision | `bfloat16` |
|
|
|
93 |
|
94 |
## Training Details
|
95 |
- Mamba-7B was trained using AWS SageMaker on 128 H100 80GB GPUs.
|
96 |
+
- Training began in March 2024 and lasted three weeks.
|
97 |
| **Hyperparameter** | **Value** |
|
98 |
|--------------------|------------|
|
99 |
| Precision | `bfloat16` |
|