tejaskamtam commited on
Commit
98a9411
1 Parent(s): b3b925b

End of training

Browse files
README.md CHANGED
@@ -3,11 +3,23 @@ license: apache-2.0
3
  base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - accuracy
8
  model-index:
9
  - name: testing
10
- results: []
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,10 +27,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # testing
17
 
18
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.7192
21
- - Accuracy: 0.6736
22
 
23
  ## Model description
24
 
 
3
  base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - datasets/all_binary_and_xe_ey_fae_counterfactual
8
  metrics:
9
  - accuracy
10
  model-index:
11
  - name: testing
12
+ results:
13
+ - task:
14
+ name: Masked Language Modeling
15
+ type: fill-mask
16
+ dataset:
17
+ name: datasets/all_binary_and_xe_ey_fae_counterfactual
18
+ type: datasets/all_binary_and_xe_ey_fae_counterfactual
19
+ metrics:
20
+ - name: Accuracy
21
+ type: accuracy
22
+ value: 0.6740440005371309
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
27
 
28
  # testing
29
 
30
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the datasets/all_binary_and_xe_ey_fae_counterfactual dataset.
31
  It achieves the following results on the evaluation set:
32
+ - Loss: 1.7135
33
+ - Accuracy: 0.6740
34
 
35
  ## Model description
36
 
all_results.json CHANGED
@@ -2,14 +2,14 @@
2
  "epoch": 0.0,
3
  "eval_accuracy": 0.6740440005371309,
4
  "eval_loss": 1.7135361433029175,
5
- "eval_runtime": 79.7989,
6
  "eval_samples": 15525,
7
- "eval_samples_per_second": 194.551,
8
- "eval_steps_per_second": 24.324,
9
  "perplexity": 5.548547286429599,
10
  "train_loss": 1.9919448852539063,
11
- "train_runtime": 161.5625,
12
  "train_samples": 124124,
13
- "train_samples_per_second": 0.99,
14
  "train_steps_per_second": 0.062
15
  }
 
2
  "epoch": 0.0,
3
  "eval_accuracy": 0.6740440005371309,
4
  "eval_loss": 1.7135361433029175,
5
+ "eval_runtime": 79.6522,
6
  "eval_samples": 15525,
7
+ "eval_samples_per_second": 194.91,
8
+ "eval_steps_per_second": 24.368,
9
  "perplexity": 5.548547286429599,
10
  "train_loss": 1.9919448852539063,
11
+ "train_runtime": 161.1653,
12
  "train_samples": 124124,
13
+ "train_samples_per_second": 0.993,
14
  "train_steps_per_second": 0.062
15
  }
eval_results.json CHANGED
@@ -2,9 +2,9 @@
2
  "epoch": 0.0,
3
  "eval_accuracy": 0.6740440005371309,
4
  "eval_loss": 1.7135361433029175,
5
- "eval_runtime": 79.7989,
6
  "eval_samples": 15525,
7
- "eval_samples_per_second": 194.551,
8
- "eval_steps_per_second": 24.324,
9
  "perplexity": 5.548547286429599
10
  }
 
2
  "epoch": 0.0,
3
  "eval_accuracy": 0.6740440005371309,
4
  "eval_loss": 1.7135361433029175,
5
+ "eval_runtime": 79.6522,
6
  "eval_samples": 15525,
7
+ "eval_samples_per_second": 194.91,
8
+ "eval_steps_per_second": 24.368,
9
  "perplexity": 5.548547286429599
10
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.0,
3
  "train_loss": 1.9919448852539063,
4
- "train_runtime": 161.5625,
5
  "train_samples": 124124,
6
- "train_samples_per_second": 0.99,
7
  "train_steps_per_second": 0.062
8
  }
 
1
  {
2
  "epoch": 0.0,
3
  "train_loss": 1.9919448852539063,
4
+ "train_runtime": 161.1653,
5
  "train_samples": 124124,
6
+ "train_samples_per_second": 0.993,
7
  "train_steps_per_second": 0.062
8
  }
trainer_state.json CHANGED
@@ -18,9 +18,9 @@
18
  "epoch": 0.0,
19
  "eval_accuracy": 0.6649938341745655,
20
  "eval_loss": 1.775834321975708,
21
- "eval_runtime": 78.8594,
22
- "eval_samples_per_second": 196.869,
23
- "eval_steps_per_second": 24.613,
24
  "step": 5
25
  },
26
  {
@@ -33,9 +33,9 @@
33
  "epoch": 0.0,
34
  "eval_accuracy": 0.6735738398990131,
35
  "eval_loss": 1.719245433807373,
36
- "eval_runtime": 79.7924,
37
- "eval_samples_per_second": 194.567,
38
- "eval_steps_per_second": 24.326,
39
  "step": 10
40
  },
41
  {
@@ -43,8 +43,8 @@
43
  "step": 10,
44
  "total_flos": 42112770048000.0,
45
  "train_loss": 1.9919448852539063,
46
- "train_runtime": 161.5625,
47
- "train_samples_per_second": 0.99,
48
  "train_steps_per_second": 0.062
49
  }
50
  ],
 
18
  "epoch": 0.0,
19
  "eval_accuracy": 0.6649938341745655,
20
  "eval_loss": 1.775834321975708,
21
+ "eval_runtime": 78.6312,
22
+ "eval_samples_per_second": 197.441,
23
+ "eval_steps_per_second": 24.685,
24
  "step": 5
25
  },
26
  {
 
33
  "epoch": 0.0,
34
  "eval_accuracy": 0.6735738398990131,
35
  "eval_loss": 1.719245433807373,
36
+ "eval_runtime": 79.5982,
37
+ "eval_samples_per_second": 195.042,
38
+ "eval_steps_per_second": 24.385,
39
  "step": 10
40
  },
41
  {
 
43
  "step": 10,
44
  "total_flos": 42112770048000.0,
45
  "train_loss": 1.9919448852539063,
46
+ "train_runtime": 161.1653,
47
+ "train_samples_per_second": 0.993,
48
  "train_steps_per_second": 0.062
49
  }
50
  ],