bhaskara / README.md
mattf1n's picture
Update README.md
3da0e67
---
license: apache-2.0
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: output
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# output
## Model description
This model is a fine-tuned version of [EleutherAI/gpt-neo-2.7B](https://huggingface.co/EleutherAI/gpt-neo-2.7B) on the Lila-IID-train/dev set from the [Lila dataset](https://github.com/allenai/Lila).
## Intended uses & limitations
If you use this model, please cite our work.
```
@INPROCEEDINGS{Mishra2022Lila,
author = {
Swaroop Mishra
and Matthew Finlayson
and Pan Lu
and Leonard Tang
and Sean Welleck
and Chitta Baral
and Tanmay Rajpurohit
and Oyvind Tafjord
and Ashish Sabharwal
and Peter Clark
and Ashwin Kalyan},
title = {Lila: A Unified Benchmark for Mathematical Reasoning},
booktitle = {Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP)},
year = {2022}
}
```
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- distributed_type: multi-GPU
- num_devices: 2
- total_train_batch_size: 8
- total_eval_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10.0
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| No log | 0.06 | 100 | 0.7930 | 0.8214 |
| No log | 0.11 | 200 | 0.7544 | 0.8290 |
| No log | 0.17 | 300 | 0.7358 | 0.8328 |
| No log | 0.23 | 400 | 0.7192 | 0.8357 |
| 0.8156 | 0.28 | 500 | 0.7012 | 0.8397 |
| 0.8156 | 0.34 | 600 | 0.6904 | 0.8419 |
| 0.8156 | 0.4 | 700 | 0.6802 | 0.8440 |
| 0.8156 | 0.45 | 800 | 0.6670 | 0.8465 |
| 0.8156 | 0.51 | 900 | 0.6572 | 0.8486 |
| 0.7219 | 0.57 | 1000 | 0.6499 | 0.8500 |
| 0.7219 | 0.62 | 1100 | 0.6411 | 0.8522 |
| 0.7219 | 0.68 | 1200 | 0.6343 | 0.8537 |
| 0.7219 | 0.74 | 1300 | 0.6299 | 0.8546 |
| 0.7219 | 0.79 | 1400 | 0.6221 | 0.8561 |
| 0.662 | 0.85 | 1500 | 0.6157 | 0.8574 |
| 0.662 | 0.91 | 1600 | 0.6138 | 0.8579 |
| 0.662 | 0.96 | 1700 | 0.6055 | 0.8595 |
| 0.662 | 1.02 | 1800 | 0.6143 | 0.8598 |
| 0.662 | 1.08 | 1900 | 0.6191 | 0.8599 |
| 0.5707 | 1.14 | 2000 | 0.6118 | 0.8607 |
| 0.5707 | 1.19 | 2100 | 0.6123 | 0.8611 |
| 0.5707 | 1.25 | 2200 | 0.6089 | 0.8617 |
| 0.5707 | 1.31 | 2300 | 0.6064 | 0.8619 |
| 0.5707 | 1.36 | 2400 | 0.6079 | 0.8625 |
| 0.4923 | 1.42 | 2500 | 0.6040 | 0.8625 |
| 0.4923 | 1.48 | 2600 | 0.6030 | 0.8630 |
| 0.4923 | 1.53 | 2700 | 0.6021 | 0.8636 |
| 0.4923 | 1.59 | 2800 | 0.6001 | 0.8643 |
| 0.4923 | 1.65 | 2900 | 0.5981 | 0.8644 |
| 0.4909 | 1.7 | 3000 | 0.5942 | 0.8648 |
| 0.4909 | 1.76 | 3100 | 0.5918 | 0.8650 |
| 0.4909 | 1.82 | 3200 | 0.5923 | 0.8659 |
| 0.4909 | 1.87 | 3300 | 0.5884 | 0.8664 |
| 0.4909 | 1.93 | 3400 | 0.5884 | 0.8663 |
| 0.4964 | 1.99 | 3500 | 0.5903 | 0.8669 |
| 0.4964 | 2.04 | 3600 | 0.6421 | 0.8655 |
| 0.4964 | 2.1 | 3700 | 0.6401 | 0.8651 |
| 0.4964 | 2.16 | 3800 | 0.6411 | 0.8649 |
| 0.4964 | 2.21 | 3900 | 0.6387 | 0.8645 |
| 0.345 | 2.27 | 4000 | 0.6362 | 0.8654 |
| 0.345 | 2.33 | 4100 | 0.6362 | 0.8654 |
| 0.345 | 2.38 | 4200 | 0.6362 | 0.8654 |
| 0.345 | 2.44 | 4300 | 0.6357 | 0.8655 |
| 0.345 | 2.5 | 4400 | 0.6362 | 0.8656 |
| 0.3463 | 2.55 | 4500 | 0.6377 | 0.8658 |
| 0.3463 | 2.61 | 4600 | 0.6357 | 0.8660 |
| 0.3463 | 2.67 | 4700 | 0.6294 | 0.8665 |
| 0.3463 | 2.72 | 4800 | 0.6333 | 0.8665 |
| 0.3463 | 2.78 | 4900 | 0.6362 | 0.8662 |
| 0.3508 | 2.84 | 5000 | 0.6357 | 0.8666 |
| 0.3508 | 2.89 | 5100 | 0.6299 | 0.8673 |
| 0.3508 | 2.95 | 5200 | 0.6313 | 0.8668 |
| 0.3508 | 3.01 | 5300 | 0.7188 | 0.8646 |
| 0.3508 | 3.06 | 5400 | 0.7017 | 0.8656 |
| 0.295 | 3.12 | 5500 | 0.6982 | 0.8653 |
| 0.295 | 3.18 | 5600 | 0.7031 | 0.8655 |
| 0.295 | 3.23 | 5700 | 0.6992 | 0.8651 |
| 0.295 | 3.29 | 5800 | 0.6997 | 0.8653 |
| 0.295 | 3.35 | 5900 | 0.7041 | 0.8651 |
| 0.2348 | 3.41 | 6000 | 0.7075 | 0.8649 |
| 0.2348 | 3.46 | 6100 | 0.6992 | 0.8650 |
| 0.2348 | 3.52 | 6200 | 0.7065 | 0.8647 |
| 0.2348 | 3.58 | 6300 | 0.6997 | 0.8652 |
| 0.2348 | 3.63 | 6400 | 0.7026 | 0.8651 |
| 0.2411 | 3.69 | 6500 | 0.7046 | 0.8656 |
| 0.2411 | 3.75 | 6600 | 0.7007 | 0.8655 |
| 0.2411 | 3.8 | 6700 | 0.7026 | 0.8651 |
| 0.2411 | 3.86 | 6800 | 0.7031 | 0.8655 |
| 0.2411 | 3.92 | 6900 | 0.7012 | 0.8658 |
| 0.251 | 3.97 | 7000 | 0.7051 | 0.8656 |
| 0.251 | 4.03 | 7100 | 0.7607 | 0.8650 |
| 0.251 | 4.09 | 7200 | 0.7632 | 0.8656 |
| 0.251 | 4.14 | 7300 | 0.7588 | 0.8655 |
| 0.251 | 4.2 | 7400 | 0.7578 | 0.8651 |
| 0.1797 | 4.26 | 7500 | 0.7710 | 0.8645 |
| 0.1797 | 4.31 | 7600 | 0.7627 | 0.8648 |
| 0.1797 | 4.37 | 7700 | 0.7583 | 0.8650 |
| 0.1797 | 4.43 | 7800 | 0.7646 | 0.8649 |
| 0.1797 | 4.48 | 7900 | 0.7598 | 0.8646 |
| 0.1784 | 4.54 | 8000 | 0.7656 | 0.8650 |
| 0.1784 | 4.6 | 8100 | 0.7617 | 0.8648 |
| 0.1784 | 4.65 | 8200 | 0.7573 | 0.8651 |
| 0.1784 | 4.71 | 8300 | 0.7671 | 0.8648 |
| 0.1784 | 4.77 | 8400 | 0.7563 | 0.8651 |
| 0.1827 | 4.82 | 8500 | 0.7651 | 0.8649 |
| 0.1827 | 4.88 | 8600 | 0.7637 | 0.8650 |
| 0.1827 | 4.94 | 8700 | 0.7607 | 0.8654 |
| 0.1827 | 4.99 | 8800 | 0.7607 | 0.8650 |
| 0.1827 | 5.05 | 8900 | 0.8149 | 0.8646 |
| 0.167 | 5.11 | 9000 | 0.8081 | 0.8648 |
| 0.167 | 5.16 | 9100 | 0.8184 | 0.8644 |
| 0.167 | 5.22 | 9200 | 0.8140 | 0.8647 |
| 0.167 | 5.28 | 9300 | 0.8169 | 0.8644 |
| 0.167 | 5.33 | 9400 | 0.8120 | 0.8645 |
| 0.1371 | 5.39 | 9500 | 0.8154 | 0.8643 |
| 0.1371 | 5.45 | 9600 | 0.8179 | 0.8642 |
| 0.1371 | 5.51 | 9700 | 0.8154 | 0.8643 |
| 0.1371 | 5.56 | 9800 | 0.8120 | 0.8645 |
| 0.1371 | 5.62 | 9900 | 0.8110 | 0.8650 |
| 0.1425 | 5.68 | 10000 | 0.8159 | 0.8645 |
| 0.1425 | 5.73 | 10100 | 0.8174 | 0.8646 |
| 0.1425 | 5.79 | 10200 | 0.8159 | 0.8649 |
| 0.1425 | 5.85 | 10300 | 0.8110 | 0.8639 |
| 0.1425 | 5.9 | 10400 | 0.8135 | 0.8645 |
| 0.1505 | 5.96 | 10500 | 0.8140 | 0.8642 |
| 0.1505 | 6.02 | 10600 | 0.8628 | 0.8640 |
| 0.1505 | 6.07 | 10700 | 0.8540 | 0.8644 |
| 0.1505 | 6.13 | 10800 | 0.8530 | 0.8642 |
| 0.1505 | 6.19 | 10900 | 0.8560 | 0.8647 |
| 0.1086 | 6.24 | 11000 | 0.8555 | 0.8649 |
| 0.1086 | 6.3 | 11100 | 0.8604 | 0.8644 |
| 0.1086 | 6.36 | 11200 | 0.8569 | 0.8642 |
| 0.1086 | 6.41 | 11300 | 0.8530 | 0.8639 |
| 0.1086 | 6.47 | 11400 | 0.8589 | 0.8643 |
| 0.1076 | 6.53 | 11500 | 0.8525 | 0.8639 |
| 0.1076 | 6.58 | 11600 | 0.8579 | 0.8640 |
| 0.1076 | 6.64 | 11700 | 0.8594 | 0.8640 |
| 0.1076 | 6.7 | 11800 | 0.8599 | 0.8643 |
| 0.1076 | 6.75 | 11900 | 0.8564 | 0.8640 |
| 0.1109 | 6.81 | 12000 | 0.8633 | 0.8640 |
| 0.1109 | 6.87 | 12100 | 0.8584 | 0.8638 |
| 0.1109 | 6.92 | 12200 | 0.8647 | 0.8636 |
| 0.1109 | 6.98 | 12300 | 0.8599 | 0.8635 |
| 0.1109 | 7.04 | 12400 | 0.8979 | 0.8632 |
| 0.1028 | 7.09 | 12500 | 0.8936 | 0.8635 |
| 0.1028 | 7.15 | 12600 | 0.9043 | 0.8637 |
| 0.1028 | 7.21 | 12700 | 0.8989 | 0.8642 |
| 0.1028 | 7.26 | 12800 | 0.8936 | 0.8642 |
| 0.1028 | 7.32 | 12900 | 0.8921 | 0.8641 |
| 0.0774 | 7.38 | 13000 | 0.8955 | 0.8634 |
| 0.0774 | 7.43 | 13100 | 0.8950 | 0.8636 |
| 0.0774 | 7.49 | 13200 | 0.8994 | 0.8635 |
| 0.0774 | 7.55 | 13300 | 0.8999 | 0.8635 |
| 0.0774 | 7.6 | 13400 | 0.8936 | 0.8631 |
| 0.0852 | 7.66 | 13500 | 0.9048 | 0.8634 |
| 0.0852 | 7.72 | 13600 | 0.8960 | 0.8632 |
| 0.0852 | 7.78 | 13700 | 0.9023 | 0.8635 |
| 0.0852 | 7.83 | 13800 | 0.8984 | 0.8638 |
| 0.0852 | 7.89 | 13900 | 0.9019 | 0.8635 |
| 0.0879 | 7.95 | 14000 | 0.9014 | 0.8634 |
| 0.0879 | 8.0 | 14100 | 0.9136 | 0.8630 |
| 0.0879 | 8.06 | 14200 | 0.9312 | 0.8639 |
| 0.0879 | 8.12 | 14300 | 0.9346 | 0.8635 |
| 0.0879 | 8.17 | 14400 | 0.9307 | 0.8635 |
| 0.0611 | 8.23 | 14500 | 0.9419 | 0.8641 |
| 0.0611 | 8.29 | 14600 | 0.9331 | 0.8631 |
| 0.0611 | 8.34 | 14700 | 0.9375 | 0.8636 |
| 0.0611 | 8.4 | 14800 | 0.9292 | 0.8626 |
| 0.0611 | 8.46 | 14900 | 0.9458 | 0.8637 |
| 0.061 | 8.51 | 15000 | 0.9336 | 0.8634 |
| 0.061 | 8.57 | 15100 | 0.9409 | 0.8630 |
| 0.061 | 8.63 | 15200 | 0.9390 | 0.8632 |
| 0.061 | 8.68 | 15300 | 0.9375 | 0.8628 |
| 0.061 | 8.74 | 15400 | 0.9365 | 0.8630 |
| 0.0646 | 8.8 | 15500 | 0.9370 | 0.8628 |
| 0.0646 | 8.85 | 15600 | 0.9355 | 0.8629 |
| 0.0646 | 8.91 | 15700 | 0.9375 | 0.8632 |
| 0.0646 | 8.97 | 15800 | 0.9390 | 0.8630 |
| 0.0646 | 9.02 | 15900 | 0.9717 | 0.8630 |
| 0.0593 | 9.08 | 16000 | 0.9673 | 0.8626 |
| 0.0593 | 9.14 | 16100 | 0.9644 | 0.8630 |
| 0.0593 | 9.19 | 16200 | 0.9624 | 0.8631 |
| 0.0593 | 9.25 | 16300 | 0.9648 | 0.8633 |
| 0.0593 | 9.31 | 16400 | 0.9673 | 0.8632 |
| 0.0415 | 9.36 | 16500 | 0.9658 | 0.8633 |
| 0.0415 | 9.42 | 16600 | 0.9688 | 0.8628 |
| 0.0415 | 9.48 | 16700 | 0.9653 | 0.8632 |
| 0.0415 | 9.53 | 16800 | 0.9658 | 0.8628 |
| 0.0415 | 9.59 | 16900 | 0.9668 | 0.8629 |
| 0.0471 | 9.65 | 17000 | 0.9604 | 0.8625 |
| 0.0471 | 9.7 | 17100 | 0.9658 | 0.8621 |
| 0.0471 | 9.76 | 17200 | 0.9731 | 0.8630 |
| 0.0471 | 9.82 | 17300 | 0.9692 | 0.8626 |
| 0.0471 | 9.88 | 17400 | 0.9673 | 0.8623 |
| 0.0528 | 9.93 | 17500 | 0.9614 | 0.8620 |
| 0.0528 | 9.99 | 17600 | 0.9697 | 0.8621 |
### Framework versions
- Transformers 4.21.0.dev0
- Pytorch 1.12.1+cu113
- Datasets 2.4.0
- Tokenizers 0.12.1