fix small typo
#2
by
MoritzLaurer
HF staff
- opened
README.md
CHANGED
@@ -107,7 +107,7 @@ Following this, we perform instruction fine-tuning on [The Cauldron](https://hug
|
|
107 |
- [atlas-math-sets](https://huggingface.co/datasets/AtlasUnified/atlas-math-sets)
|
108 |
- [goat](https://huggingface.co/datasets/tiedong/goat)
|
109 |
|
110 |
-
We use Lora to train the parameters initialized from pre-trained backbones and full fine-tuning for newly initialized parameters (modality connector), as we find this strategy to be more stable as
|
111 |
|
112 |
More details (training procedure, data selection, hyper-parameters, etc.) along with lessons learned from our ablations will be available in an upcoming technical report.
|
113 |
|
|
|
107 |
- [atlas-math-sets](https://huggingface.co/datasets/AtlasUnified/atlas-math-sets)
|
108 |
- [goat](https://huggingface.co/datasets/tiedong/goat)
|
109 |
|
110 |
+
We use Lora to train the parameters initialized from pre-trained backbones and full fine-tuning for newly initialized parameters (modality connector), as we find this strategy to be more stable as well as more computationally efficient.
|
111 |
|
112 |
More details (training procedure, data selection, hyper-parameters, etc.) along with lessons learned from our ablations will be available in an upcoming technical report.
|
113 |
|