marinone94 commited on
Commit
fbe783d
β€’
1 Parent(s): 4591caf

Training in progress, step 50

Browse files
{checkpoint-40 β†’ checkpoint-100}/config.json RENAMED
File without changes
{checkpoint-40 β†’ checkpoint-100}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6dd2f6e11f8931247aaf820f2fb0419ac8040fa59b8ef05119421d1d540f17c
3
  size 2490337361
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1722a491ea58423692abe6cde8aac0703fa17d8346a9ae868114f2509e04c97
3
  size 2490337361
{checkpoint-40 β†’ checkpoint-100}/preprocessor_config.json RENAMED
File without changes
{checkpoint-60 β†’ checkpoint-100}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed09bae03da69e6e7d54c45ff801a7cd8e65f5bb600585b021024798e5fafa4a
3
  size 1262063089
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7998265fd0c18bea1f246ca75cca77edd2def33646432a9437b4d839e97eab3d
3
  size 1262063089
{checkpoint-60 β†’ checkpoint-100}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d5db96c9803a1c8f4051b38118bfa8389597dfc7915de251da9a5e1c69d3bc9
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24a8cdd3a3488a3315f9617b64f16798a9e655a6d49b281bfbe024d5e3f2a880
3
+ size 14503
{checkpoint-60 β†’ checkpoint-100}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81ed4ec308038905cbb975b9b2ae3c7852901f7a7bb341d7393b61de64cbb13f
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f28dfdeb34cb1b60f5d9c9b44dc5523bf2866ea557da6b58b27104a0030e0b3
3
  size 559
{checkpoint-40 β†’ checkpoint-100}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c14d79112c3d0bf1ec781a02506228c0a8892e541191602a074e77b077eb126d
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6af49a56cc0b3dfaf6c4d29d3b9ab21c908c0ad2d95e41f19e6b8101ed863a39
3
  size 623
{checkpoint-60 β†’ checkpoint-100}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.7947019867549668,
5
- "global_step": 60,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -69,11 +69,53 @@
69
  "eval_steps_per_second": 0.814,
70
  "eval_wer": 1.0,
71
  "step": 60
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
  }
73
  ],
74
  "max_steps": 750,
75
  "num_train_epochs": 10,
76
- "total_flos": 1.21965615465984e+18,
77
  "trial_name": null,
78
  "trial_params": null
79
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.3311258278145695,
5
+ "global_step": 100,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
69
  "eval_steps_per_second": 0.814,
70
  "eval_wer": 1.0,
71
  "step": 60
72
+ },
73
+ {
74
+ "epoch": 0.93,
75
+ "learning_rate": 6.959183673469387e-05,
76
+ "loss": 3.7607,
77
+ "step": 70
78
+ },
79
+ {
80
+ "epoch": 1.07,
81
+ "learning_rate": 6.857142857142857e-05,
82
+ "loss": 3.638,
83
+ "step": 80
84
+ },
85
+ {
86
+ "epoch": 1.07,
87
+ "eval_loss": 3.4979612827301025,
88
+ "eval_runtime": 5.0758,
89
+ "eval_samples_per_second": 19.307,
90
+ "eval_steps_per_second": 0.788,
91
+ "eval_wer": 1.0,
92
+ "step": 80
93
+ },
94
+ {
95
+ "epoch": 1.2,
96
+ "learning_rate": 6.755102040816325e-05,
97
+ "loss": 3.4601,
98
+ "step": 90
99
+ },
100
+ {
101
+ "epoch": 1.33,
102
+ "learning_rate": 6.653061224489796e-05,
103
+ "loss": 3.4138,
104
+ "step": 100
105
+ },
106
+ {
107
+ "epoch": 1.33,
108
+ "eval_loss": 3.489065170288086,
109
+ "eval_runtime": 4.9074,
110
+ "eval_samples_per_second": 19.97,
111
+ "eval_steps_per_second": 0.815,
112
+ "eval_wer": 1.0,
113
+ "step": 100
114
  }
115
  ],
116
  "max_steps": 750,
117
  "num_train_epochs": 10,
118
+ "total_flos": 2.00880445675488e+18,
119
  "trial_name": null,
120
  "trial_params": null
121
  }
{checkpoint-40 β†’ checkpoint-100}/training_args.bin RENAMED
File without changes
checkpoint-40/trainer_state.json DELETED
@@ -1,58 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 0.5298013245033113,
5
- "global_step": 40,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.13,
12
- "learning_rate": 4.9999999999999996e-05,
13
- "loss": 13.2048,
14
- "step": 10
15
- },
16
- {
17
- "epoch": 0.26,
18
- "learning_rate": 7.46938775510204e-05,
19
- "loss": 26.5657,
20
- "step": 20
21
- },
22
- {
23
- "epoch": 0.26,
24
- "eval_loss": 17.199691772460938,
25
- "eval_runtime": 5.248,
26
- "eval_samples_per_second": 18.674,
27
- "eval_steps_per_second": 0.762,
28
- "eval_wer": 1.0,
29
- "step": 20
30
- },
31
- {
32
- "epoch": 0.4,
33
- "learning_rate": 7.36734693877551e-05,
34
- "loss": 8.0091,
35
- "step": 30
36
- },
37
- {
38
- "epoch": 0.53,
39
- "learning_rate": 7.265306122448979e-05,
40
- "loss": 7.6332,
41
- "step": 40
42
- },
43
- {
44
- "epoch": 0.53,
45
- "eval_loss": 4.783994197845459,
46
- "eval_runtime": 4.9818,
47
- "eval_samples_per_second": 19.672,
48
- "eval_steps_per_second": 0.803,
49
- "eval_wer": 1.0,
50
- "step": 40
51
- }
52
- ],
53
- "max_steps": 750,
54
- "num_train_epochs": 10,
55
- "total_flos": 8.2206860007168e+17,
56
- "trial_name": null,
57
- "trial_params": null
58
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
{checkpoint-60 β†’ checkpoint-50}/config.json RENAMED
File without changes
{checkpoint-60 β†’ checkpoint-50}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d06df765e9a7d7ea6615404de49841893d9b9ce55d42f0bfd5486fd7105d83c4
3
  size 2490337361
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f8c3352ebf31a7c0483167400d3399b22ccc2be15a1cf307f0d154c758e9499
3
  size 2490337361
{checkpoint-60 β†’ checkpoint-50}/preprocessor_config.json RENAMED
File without changes
{checkpoint-40 β†’ checkpoint-50}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7825b23117146d129b9fbfae78e758d25f4f6e94f91b7f4977dd069a0f502f81
3
  size 1262063089
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80ebb6ebc7a95b13fda5cdc22fab5fcfb9ffdb99ca0102065a7147116a7f9f3f
3
  size 1262063089
{checkpoint-40 β†’ checkpoint-50}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9dd2cfddafbb139c577d6100cda2fcc2338327f63d6bca8b79767529c7f40e74
3
- size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe8f634109af89a7e3706d16f2fb93f152eb68502ebee688d78216511e43324
3
+ size 14503
{checkpoint-40 β†’ checkpoint-50}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf412e9944a68f26d1088ddec860c9d22492bf7207573137d3159a8ac457e674
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a7ca717969781e5bd7f66c4b808307bbfaef76d5ae72d82fe31d60dea27fb7
3
  size 559
{checkpoint-60 β†’ checkpoint-50}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d173dd827ce26f584cfa95a40de1490d99e5852c0a8f273527eda424dd24205a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5336e922700acd511fdefd5fbbe9ccf91f59901dbca8c97f53c0892943c4b7
3
  size 623
checkpoint-50/trainer_state.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.6389776357827476,
5
+ "global_step": 50,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.26,
12
+ "learning_rate": 0.0007307692307692308,
13
+ "loss": 4.2559,
14
+ "step": 20
15
+ },
16
+ {
17
+ "epoch": 0.51,
18
+ "learning_rate": 0.0007115384615384615,
19
+ "loss": 3.068,
20
+ "step": 40
21
+ },
22
+ {
23
+ "epoch": 0.64,
24
+ "eval_loss": 3.058549165725708,
25
+ "eval_runtime": 3.3353,
26
+ "eval_samples_per_second": 29.982,
27
+ "eval_steps_per_second": 1.199,
28
+ "eval_wer": 1.0,
29
+ "step": 50
30
+ }
31
+ ],
32
+ "max_steps": 780,
33
+ "num_train_epochs": 10,
34
+ "total_flos": 7.755792033779712e+17,
35
+ "trial_name": null,
36
+ "trial_params": null
37
+ }
{checkpoint-60 β†’ checkpoint-50}/training_args.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:470abf25a211ef5ec3a9821c884f2749839bee5e95f633b168d26f90c1809409
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31becb5e425437ea470512000be6c31bb502b71fe15942ff8cf0cc01eb8267a0
3
  size 3055
old_run.sh CHANGED
@@ -10,15 +10,15 @@ python old_run_asr_ctc.py \
10
  --per_device_train_batch_size="32" \
11
  --per_device_eval_batch_size="32" \
12
  --gradient_accumulation_steps="4" \
13
- --learning_rate="7.5e-5" \
14
  --max_train_samples="10000" \
15
  --max_eval_samples="100" \
16
  --warmup_ratio="0.0" \
17
  --length_column_name="input_length" \
18
  --evaluation_strategy="steps" \
19
  --save_strategy="steps" \
20
- --eval_steps="100" \
21
- --save_steps="100" \
22
  --text_column_name="sentence" \
23
  --chars_to_ignore , ? . ! \- \; \: \" β€œ % β€˜ ” οΏ½ β€” ’ … – \
24
  --logging_steps="20" \
 
10
  --per_device_train_batch_size="32" \
11
  --per_device_eval_batch_size="32" \
12
  --gradient_accumulation_steps="4" \
13
+ --learning_rate="7.5e-4" \
14
  --max_train_samples="10000" \
15
  --max_eval_samples="100" \
16
  --warmup_ratio="0.0" \
17
  --length_column_name="input_length" \
18
  --evaluation_strategy="steps" \
19
  --save_strategy="steps" \
20
+ --eval_steps="50" \
21
+ --save_steps="50" \
22
  --text_column_name="sentence" \
23
  --chars_to_ignore , ? . ! \- \; \: \" β€œ % β€˜ ” οΏ½ β€” ’ … – \
24
  --logging_steps="20" \
old_run_asr_ctc.py CHANGED
@@ -783,13 +783,17 @@ def main():
783
  # If dataset_seed is set, shuffle train
784
  if data_args.dataset_seed is not None:
785
  vectorized_datasets["train"] = vectorized_datasets["train"].shuffle(seed=data_args.dataset_seed)
786
-
787
  # TODO: Log sample of datasets in the right way (see wandb docs)
788
  pd_train = vectorized_datasets["train"].select(range(10)).to_pandas()
789
  pd_eval = vectorized_datasets["eval"].select(range(10)).to_pandas()
790
  # wandb.log({"train_sample": pd_train})
791
  # wandb.log({"eval_sample": pd_eval})
792
 
 
 
 
 
793
  # for large datasets it is advised to run the preprocessing on a
794
  # single machine first with ``args.preprocessing_only`` since there will mostly likely
795
  # be a timeout when running the script in distributed mode.
@@ -806,9 +810,12 @@ def main():
806
  pred.label_ids[pred.label_ids == -100] = tokenizer.pad_token_id
807
 
808
  pred_str = tokenizer.batch_decode(pred_ids)
 
809
  # we do not want to group tokens when computing the metrics
810
  label_str = tokenizer.batch_decode(pred.label_ids, group_tokens=False)
811
 
 
 
812
  metrics = {k: v.compute(predictions=pred_str, references=label_str) for k, v in eval_metrics.items()}
813
 
814
  return metrics
 
783
  # If dataset_seed is set, shuffle train
784
  if data_args.dataset_seed is not None:
785
  vectorized_datasets["train"] = vectorized_datasets["train"].shuffle(seed=data_args.dataset_seed)
786
+
787
  # TODO: Log sample of datasets in the right way (see wandb docs)
788
  pd_train = vectorized_datasets["train"].select(range(10)).to_pandas()
789
  pd_eval = vectorized_datasets["eval"].select(range(10)).to_pandas()
790
  # wandb.log({"train_sample": pd_train})
791
  # wandb.log({"eval_sample": pd_eval})
792
 
793
+ print(pd_train)
794
+ print(pd_eval)
795
+
796
+
797
  # for large datasets it is advised to run the preprocessing on a
798
  # single machine first with ``args.preprocessing_only`` since there will mostly likely
799
  # be a timeout when running the script in distributed mode.
 
810
  pred.label_ids[pred.label_ids == -100] = tokenizer.pad_token_id
811
 
812
  pred_str = tokenizer.batch_decode(pred_ids)
813
+
814
  # we do not want to group tokens when computing the metrics
815
  label_str = tokenizer.batch_decode(pred.label_ids, group_tokens=False)
816
 
817
+ print(pred_str)
818
+ print(label_str)
819
  metrics = {k: v.compute(predictions=pred_str, references=label_str) for k, v in eval_metrics.items()}
820
 
821
  return metrics
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d243142fcbd07def1347c28af05f32d03aa37c7536de513cc8503d261cf6d773
3
  size 1262063089
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80ebb6ebc7a95b13fda5cdc22fab5fcfb9ffdb99ca0102065a7147116a7f9f3f
3
  size 1262063089
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:470abf25a211ef5ec3a9821c884f2749839bee5e95f633b168d26f90c1809409
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31becb5e425437ea470512000be6c31bb502b71fe15942ff8cf0cc01eb8267a0
3
  size 3055