againeureka commited on
Commit
978c73f
1 Parent(s): a47477d

Training in progress, epoch 1

Browse files
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 5.0,
3
- "eval_accuracy": 0.6909560723514212,
4
- "eval_f1": 0.689969114020938,
5
- "eval_loss": 0.8846441507339478,
6
- "eval_precision": 0.704645887749851,
7
- "eval_recall": 0.6849033527022119,
8
- "eval_runtime": 1.5948,
9
- "eval_samples_per_second": 1213.288,
10
- "eval_steps_per_second": 75.87,
11
  "total_flos": 754719218037696.0,
12
  "train_loss": 1.1282823350694444,
13
  "train_runtime": 131.3823,
 
1
  {
2
  "epoch": 5.0,
3
+ "eval_accuracy": 0.8428940568475453,
4
+ "eval_f1": 0.8699568253701934,
5
+ "eval_loss": 0.47272029519081116,
6
+ "eval_precision": 0.8642716094407783,
7
+ "eval_recall": 0.8785123251677769,
8
+ "eval_runtime": 1.7019,
9
+ "eval_samples_per_second": 1136.932,
10
+ "eval_steps_per_second": 71.095,
11
  "total_flos": 754719218037696.0,
12
  "train_loss": 1.1282823350694444,
13
  "train_runtime": 131.3823,
eval_results.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
  "epoch": 5.0,
3
- "eval_accuracy": 0.6909560723514212,
4
- "eval_f1": 0.689969114020938,
5
- "eval_loss": 0.8846441507339478,
6
- "eval_precision": 0.704645887749851,
7
- "eval_recall": 0.6849033527022119,
8
- "eval_runtime": 1.5948,
9
- "eval_samples_per_second": 1213.288,
10
- "eval_steps_per_second": 75.87
11
  }
 
1
  {
2
  "epoch": 5.0,
3
+ "eval_accuracy": 0.8428940568475453,
4
+ "eval_f1": 0.8699568253701934,
5
+ "eval_loss": 0.47272029519081116,
6
+ "eval_precision": 0.8642716094407783,
7
+ "eval_recall": 0.8785123251677769,
8
+ "eval_runtime": 1.7019,
9
+ "eval_samples_per_second": 1136.932,
10
+ "eval_steps_per_second": 71.095
11
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15cba8d519d8ca644993eb10fe9f278357a1394e7492af7de6064a055d720c70
3
  size 442570421
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00a928a61153770e4e95bf84c7649ad429dc97517122828744b5f6602234ff90
3
  size 442570421
special_tokens_map.json CHANGED
@@ -1,7 +1,5 @@
1
  {
2
- "bos_token": "[CLS]",
3
  "cls_token": "[CLS]",
4
- "eos_token": "[SEP]",
5
  "mask_token": "[MASK]",
6
  "pad_token": "[PAD]",
7
  "sep_token": "[SEP]",
 
1
  {
 
2
  "cls_token": "[CLS]",
 
3
  "mask_token": "[MASK]",
4
  "pad_token": "[PAD]",
5
  "sep_token": "[SEP]",
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f528deccb283323c613f41f4cd7cd1ab207d533f13acf524fc27678317e2ed5c
3
- size 752097
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc88a803ad3114200710eeef2106ecff7b5ac010352201b52b17aa62e2fffb9b
3
+ size 1183633
tokenizer_config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
2
- "bos_token": "[CLS]",
3
  "clean_up_tokenization_spaces": true,
4
  "cls_token": "[CLS]",
5
  "do_basic_tokenize": true,
6
- "do_lower_case": false,
7
- "eos_token": "[SEP]",
8
  "mask_token": "[MASK]",
9
  "model_max_length": 512,
10
  "never_split": null,
 
1
  {
 
2
  "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_basic_tokenize": true,
5
+ "do_lower_case": true,
 
6
  "mask_token": "[MASK]",
7
  "model_max_length": 512,
8
  "never_split": null,
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff