albertusgeyser commited on
Commit
76fbf36
1 Parent(s): fc2f352

Training in progress, step 20, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "v_proj",
27
- "q_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
  ],
29
  "task_type": null,
30
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc753107f460eebedcab37a978dc8f4e376ea7b0a7a6245079a1bec869b47db3
3
  size 14176064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b41aefecebf602e8b3eb83165cb11ae91e063ef795fcd5382f52b0a4834c8eaf
3
  size 14176064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63f1cc00fb95d4564acd83c7d51ff7e80f513b3607c768e407fdf3c02d599d6a
3
  size 4832762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfec19d66473e683c8c6c8c02124aed3a2badb2fd840f8f887295bbe6b53fb09
3
  size 4832762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97651e5195311fa83aac8286b9c0a9068e32980aed21ebb08fafd1001eee44d6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac57d8346a3be754b4ec98f02712a7da2e74f2ce81326866ad29a9acb7908177
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df73be84f0c00f13e358d00950761edb4aec7bfdee9e97f169f67cba14d6b4b5
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94db20b168116ecb35fae58feadaa8ad67980f0f0db9d75b42e7896ad19770de
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,92 +1,28 @@
1
  {
2
- "best_metric": 63.72093023255814,
3
- "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-40",
4
- "epoch": 99.01,
5
  "eval_steps": 20,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 19.01,
13
- "grad_norm": 1.7477251291275024,
14
  "learning_rate": 0.0001,
15
- "loss": 4.6928,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 19.01,
20
- "eval_loss": 3.541473388671875,
21
- "eval_runtime": 16.3765,
22
- "eval_samples_per_second": 1.221,
23
- "eval_steps_per_second": 0.061,
24
- "eval_wer": 66.04651162790698,
25
- "step": 20
26
- },
27
- {
28
- "epoch": 39.01,
29
- "grad_norm": 1.1091852188110352,
30
- "learning_rate": 0.0001,
31
- "loss": 2.4177,
32
- "step": 40
33
- },
34
- {
35
- "epoch": 39.01,
36
- "eval_loss": 2.713858127593994,
37
- "eval_runtime": 15.3439,
38
- "eval_samples_per_second": 1.303,
39
  "eval_steps_per_second": 0.065,
40
- "eval_wer": 63.72093023255814,
41
- "step": 40
42
- },
43
- {
44
- "epoch": 59.01,
45
- "grad_norm": 0.5135031342506409,
46
- "learning_rate": 0.0001,
47
- "loss": 1.1967,
48
- "step": 60
49
- },
50
- {
51
- "epoch": 59.01,
52
- "eval_loss": 2.546757459640503,
53
- "eval_runtime": 16.0308,
54
- "eval_samples_per_second": 1.248,
55
- "eval_steps_per_second": 0.062,
56
- "eval_wer": 66.97674418604652,
57
- "step": 60
58
- },
59
- {
60
- "epoch": 79.01,
61
- "grad_norm": 0.16271141171455383,
62
- "learning_rate": 0.0001,
63
- "loss": 0.7102,
64
- "step": 80
65
- },
66
- {
67
- "epoch": 79.01,
68
- "eval_loss": 2.530426502227783,
69
- "eval_runtime": 45.1139,
70
- "eval_samples_per_second": 0.443,
71
- "eval_steps_per_second": 0.022,
72
- "eval_wer": 110.23255813953487,
73
- "step": 80
74
- },
75
- {
76
- "epoch": 99.01,
77
- "grad_norm": 0.1356694996356964,
78
- "learning_rate": 0.0001,
79
- "loss": 0.609,
80
- "step": 100
81
- },
82
- {
83
- "epoch": 99.01,
84
- "eval_loss": 2.4960544109344482,
85
- "eval_runtime": 45.3429,
86
- "eval_samples_per_second": 0.441,
87
- "eval_steps_per_second": 0.022,
88
- "eval_wer": 164.65116279069767,
89
- "step": 100
90
  }
91
  ],
92
  "logging_steps": 20,
@@ -101,12 +37,12 @@
101
  "should_evaluate": false,
102
  "should_log": false,
103
  "should_save": true,
104
- "should_training_stop": true
105
  },
106
  "attributes": {}
107
  }
108
  },
109
- "total_flos": 5.8736295936e+17,
110
  "train_batch_size": 32,
111
  "trial_name": null,
112
  "trial_params": null
 
1
  {
2
+ "best_metric": 66.51162790697674,
3
+ "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-20",
4
+ "epoch": 19.01,
5
  "eval_steps": 20,
6
+ "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 19.01,
13
+ "grad_norm": 1.7188540697097778,
14
  "learning_rate": 0.0001,
15
+ "loss": 4.6963,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 19.01,
20
+ "eval_loss": 3.55057692527771,
21
+ "eval_runtime": 15.3063,
22
+ "eval_samples_per_second": 1.307,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "eval_steps_per_second": 0.065,
24
+ "eval_wer": 66.51162790697674,
25
+ "step": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  }
27
  ],
28
  "logging_steps": 20,
 
37
  "should_evaluate": false,
38
  "should_log": false,
39
  "should_save": true,
40
+ "should_training_stop": false
41
  },
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 1.17472591872e+17,
46
  "train_batch_size": 32,
47
  "trial_name": null,
48
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e9c0f990cd62c4a314194f8df86f013fa125a968b85201927e1c9b01694c489
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f857dc0a6b1c734e5acd997e0dfa2bd93ce1b14c193a41b1ec395d28b179cc2
3
  size 5368