Update README.md
Browse files
README.md
CHANGED
@@ -43,15 +43,13 @@ print(tokenizer.decode(outputs[0]))
|
|
43 |
|
44 |
Model evaluation metrics and results.
|
45 |
|
46 |
-
| Benchmark
|
47 |
-
|
48 |
-
| [GSM8K](https://arxiv.org/abs/2110.14168)
|
49 |
|
50 |
|
51 |
## Model Training Details
|
52 |
|
53 |
-
sp0_2ep_lr3e-5_bs32_warmup20ba
|
54 |
-
|
55 |
This model was obtained by fine-tuning the [dense Llama 2 7B](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the [GSM8k](https://huggingface.co/datasets/openai/gsm8k) dataset.
|
56 |
Fine-tuning was performed for 2 epochs with batch-size of 32, with linearly decaying learning-rate from initial value of 3e-5 and warm-up phase of 20 steps.
|
57 |
|
|
|
43 |
|
44 |
Model evaluation metrics and results.
|
45 |
|
46 |
+
| Benchmark | Metric | Llama-2-7b-gsm8k |
|
47 |
+
|:----:|:----:|:----:|
|
48 |
+
| [GSM8K](https://arxiv.org/abs/2110.14168) | 0-shot | 35.5% |
|
49 |
|
50 |
|
51 |
## Model Training Details
|
52 |
|
|
|
|
|
53 |
This model was obtained by fine-tuning the [dense Llama 2 7B](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the [GSM8k](https://huggingface.co/datasets/openai/gsm8k) dataset.
|
54 |
Fine-tuning was performed for 2 epochs with batch-size of 32, with linearly decaying learning-rate from initial value of 3e-5 and warm-up phase of 20 steps.
|
55 |
|