ekurtic commited on
Commit
1d435c9
1 Parent(s): e599123

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -43,15 +43,13 @@ print(tokenizer.decode(outputs[0]))
43
 
44
  Model evaluation metrics and results.
45
 
46
- | Benchmark | Metric | Llama-2-7b-gsm8k |
47
- |------------------------------------------------|---------------|-------------|
48
- | [GSM8K](https://arxiv.org/abs/2110.14168) | 0-shot | 35.5% |
49
 
50
 
51
  ## Model Training Details
52
 
53
- sp0_2ep_lr3e-5_bs32_warmup20ba
54
-
55
  This model was obtained by fine-tuning the [dense Llama 2 7B](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the [GSM8k](https://huggingface.co/datasets/openai/gsm8k) dataset.
56
  Fine-tuning was performed for 2 epochs with batch-size of 32, with linearly decaying learning-rate from initial value of 3e-5 and warm-up phase of 20 steps.
57
 
 
43
 
44
  Model evaluation metrics and results.
45
 
46
+ | Benchmark | Metric | Llama-2-7b-gsm8k |
47
+ |:----:|:----:|:----:|
48
+ | [GSM8K](https://arxiv.org/abs/2110.14168) | 0-shot | 35.5% |
49
 
50
 
51
  ## Model Training Details
52
 
 
 
53
  This model was obtained by fine-tuning the [dense Llama 2 7B](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the [GSM8k](https://huggingface.co/datasets/openai/gsm8k) dataset.
54
  Fine-tuning was performed for 2 epochs with batch-size of 32, with linearly decaying learning-rate from initial value of 3e-5 and warm-up phase of 20 steps.
55