mamlong34 commited on
Commit
b3698d2
1 Parent(s): 96fbae0

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -13
README.md CHANGED
@@ -3,7 +3,7 @@ license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
- - cosmos_qa
7
  metrics:
8
  - accuracy
9
  model-index:
@@ -16,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # t5_base_race_cosmos_qa
18
 
19
- This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the cosmos_qa dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.5038
22
- - Accuracy: 0.7176
23
 
24
  ## Model description
25
 
@@ -39,11 +39,9 @@ More information needed
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 0.0001
42
- - train_batch_size: 2
43
  - eval_batch_size: 16
44
  - seed: 42
45
- - gradient_accumulation_steps: 4
46
- - total_train_batch_size: 8
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 100
@@ -51,16 +49,16 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
55
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
56
- | 0.4408 | 1.0 | 3157 | 0.4825 | 0.6315 |
57
- | 0.3166 | 2.0 | 6314 | 0.4108 | 0.7109 |
58
- | 0.2 | 3.0 | 9471 | 0.5038 | 0.7176 |
59
 
60
 
61
  ### Framework versions
62
 
63
  - Transformers 4.11.3
64
- - Pytorch 1.9.1
65
  - Datasets 1.12.1
66
  - Tokenizers 0.10.3
 
3
  tags:
4
  - generated_from_trainer
5
  datasets:
6
+ - race
7
  metrics:
8
  - accuracy
9
  model-index:
 
16
 
17
  # t5_base_race_cosmos_qa
18
 
19
+ This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the race dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4414
22
+ - Accuracy: 0.7424
23
 
24
  ## Model description
25
 
 
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 0.0001
42
+ - train_batch_size: 8
43
  - eval_batch_size: 16
44
  - seed: 42
 
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 100
 
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|
54
+ | 0.4355 | 1.0 | 10984 | 0.3910 | 0.7072 |
55
+ | 0.3233 | 2.0 | 21968 | 0.3833 | 0.7321 |
56
+ | 0.229 | 3.0 | 32952 | 0.4414 | 0.7424 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.11.3
62
+ - Pytorch 1.9.0
63
  - Datasets 1.12.1
64
  - Tokenizers 0.10.3