ChrisZeng commited on
Commit
979a368
1 Parent(s): c7bc789

Training in progress, epoch 2

Browse files
checkpoint-405/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e77c8a2a5fad17d46427af1d8a570e4d5fd98c6b9ed138860d70a3bfe45c497a
3
  size 1115518581
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d06c66d25a583e1552589360dd807828fa2e9ef8fe064adf663c2930cd10750
3
  size 1115518581
checkpoint-405/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b214a195ba865a570b4c0ca7858a07d3cdf972c6a40b69118e204806f82334d3
3
  size 557982265
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61ada5f16a59af821e82557d0cc1154b9a8317ada4b526eedac4e607d1c6f078
3
  size 557982265
checkpoint-405/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:361b412c6099209dd7c32c3cf59809aa5d7b27574edb1597b9b27827ea576a15
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89be3ff2dd97ecabb1e4c4328f68780edb24a6834a6485d041b06fde96f5ac91
3
  size 623
checkpoint-405/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.1801992505788803,
3
  "best_model_checkpoint": "outputs/bart-base-detox/checkpoint-405",
4
  "epoch": 2.9953917050691246,
5
  "global_step": 405,
@@ -9,44 +9,44 @@
9
  "log_history": [
10
  {
11
  "epoch": 1.0,
12
- "learning_rate": 9e-05,
13
- "loss": 0.3171,
14
  "step": 135
15
  },
16
  {
17
  "epoch": 1.0,
18
- "eval_loss": 0.20179320871829987,
19
- "eval_runtime": 11.3072,
20
- "eval_samples_per_second": 274.16,
21
- "eval_steps_per_second": 34.314,
22
  "step": 135
23
  },
24
  {
25
  "epoch": 2.0,
26
- "learning_rate": 8e-05,
27
- "loss": 0.1852,
28
  "step": 270
29
  },
30
  {
31
  "epoch": 2.0,
32
- "eval_loss": 0.18728719651699066,
33
- "eval_runtime": 11.1563,
34
- "eval_samples_per_second": 277.869,
35
- "eval_steps_per_second": 34.778,
36
  "step": 270
37
  },
38
  {
39
  "epoch": 3.0,
40
- "learning_rate": 7e-05,
41
- "loss": 0.1559,
42
  "step": 405
43
  },
44
  {
45
  "epoch": 3.0,
46
- "eval_loss": 0.1801992505788803,
47
- "eval_runtime": 11.3199,
48
- "eval_samples_per_second": 273.853,
49
- "eval_steps_per_second": 34.276,
50
  "step": 405
51
  }
52
  ],
 
1
  {
2
+ "best_metric": 0.19929151237010956,
3
  "best_model_checkpoint": "outputs/bart-base-detox/checkpoint-405",
4
  "epoch": 2.9953917050691246,
5
  "global_step": 405,
 
9
  "log_history": [
10
  {
11
  "epoch": 1.0,
12
+ "learning_rate": 9e-06,
13
+ "loss": 0.5633,
14
  "step": 135
15
  },
16
  {
17
  "epoch": 1.0,
18
+ "eval_loss": 0.25235414505004883,
19
+ "eval_runtime": 12.2006,
20
+ "eval_samples_per_second": 254.085,
21
+ "eval_steps_per_second": 31.802,
22
  "step": 135
23
  },
24
  {
25
  "epoch": 2.0,
26
+ "learning_rate": 8.000000000000001e-06,
27
+ "loss": 0.2589,
28
  "step": 270
29
  },
30
  {
31
  "epoch": 2.0,
32
+ "eval_loss": 0.21927106380462646,
33
+ "eval_runtime": 11.2792,
34
+ "eval_samples_per_second": 274.842,
35
+ "eval_steps_per_second": 34.4,
36
  "step": 270
37
  },
38
  {
39
  "epoch": 3.0,
40
+ "learning_rate": 7e-06,
41
+ "loss": 0.2307,
42
  "step": 405
43
  },
44
  {
45
  "epoch": 3.0,
46
+ "eval_loss": 0.19929151237010956,
47
+ "eval_runtime": 11.2313,
48
+ "eval_samples_per_second": 276.016,
49
+ "eval_steps_per_second": 34.546,
50
  "step": 405
51
  }
52
  ],
checkpoint-405/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6779c0d7bfc5cc8b7f502411db449902ffdb3fd6d5cc4758f90defaa037cdfa
3
  size 3183
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd8019ba81e4d8cd7e878f72b37c14bf5f28f7f6b5acd2cf343bb5ba08c77b6a
3
  size 3183
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1328007474fa4e55660b5739283edac44d62b1a9128017d3afdee46e71e1d830
3
  size 557982265
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61ada5f16a59af821e82557d0cc1154b9a8317ada4b526eedac4e607d1c6f078
3
  size 557982265