FILALIHicham commited on
Commit
ff90c03
·
verified ·
1 Parent(s): 47ce2a2

Training in progress, step 5, checkpoint

Browse files
last-checkpoint/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/ubuntu/whisper-large-v3-praxy-fr",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForConditionalGeneration"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": [
11
+ 220,
12
+ 50257
13
+ ],
14
+ "bos_token_id": 50257,
15
+ "classifier_proj_size": 256,
16
+ "d_model": 1280,
17
+ "decoder_attention_heads": 20,
18
+ "decoder_ffn_dim": 5120,
19
+ "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 32,
21
+ "decoder_start_token_id": 50258,
22
+ "dropout": 0.0,
23
+ "encoder_attention_heads": 20,
24
+ "encoder_ffn_dim": 5120,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 32,
27
+ "eos_token_id": 50257,
28
+ "init_std": 0.02,
29
+ "is_encoder_decoder": true,
30
+ "mask_feature_length": 10,
31
+ "mask_feature_min_masks": 0,
32
+ "mask_feature_prob": 0.05,
33
+ "mask_time_length": 10,
34
+ "mask_time_min_masks": 2,
35
+ "mask_time_prob": 0.05,
36
+ "max_length": 448,
37
+ "max_source_positions": 1500,
38
+ "max_target_positions": 448,
39
+ "median_filter_width": 7,
40
+ "model_type": "whisper",
41
+ "num_hidden_layers": 32,
42
+ "num_mel_bins": 128,
43
+ "pad_token_id": 50256,
44
+ "scale_embedding": false,
45
+ "torch_dtype": "float32",
46
+ "transformers_version": "4.44.2",
47
+ "use_cache": true,
48
+ "use_weighted_layer_sum": false,
49
+ "vocab_size": 51866
50
+ }
last-checkpoint/generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "max_length": 448,
10
+ "pad_token_id": 50256,
11
+ "transformers_version": "4.44.2"
12
+ }
last-checkpoint/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae1df5dd89685564d18b95f94eafc56ee767d829b2656f10dd98a3c86204f8e
3
+ size 4993448880
last-checkpoint/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0417511e1068133a9847e813d1b44752eacb893edce1e5f9ed1e3fa173f283
3
+ size 1180663192
last-checkpoint/model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11b175638d183fff91e19e6e715f3cdfb271e97fd6b0b8d0cd2cbca94bae719e
3
+ size 12333660476
last-checkpoint/preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 128,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
last-checkpoint/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26420d0b671d327bab461c54783c75b8b78d7f1f0cdc89b12a2714ea41a32456
3
+ size 15984
last-checkpoint/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14581072dd1c58a34c134e320872f680f8eef49fbea678b198d574e8a3fe6659
3
+ size 15984
last-checkpoint/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ac47944303985f0095a7b87e35a5f146edd5c17bbe764630e646901a717b04b
3
+ size 15984
last-checkpoint/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d293ed67e54ddd72c9dbd4b89d42aaec630f294695c73a65ae50c6cc2f4c5cef
3
+ size 15984
last-checkpoint/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb73775b33aa7ed953e15ed92187b06c59b5e0ae3d7bc198ad085927dd0c1daa
3
+ size 15984
last-checkpoint/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9b1eb7016305b6071d0f8d663d5e0a346f6637993b668bf99c9a7ec9c83b25
3
+ size 15984
last-checkpoint/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a4b7b9bc852a1f39c6cea3e347663af0ee6245b784841714ef90f8c845820f4
3
+ size 15984
last-checkpoint/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a88ffde7826ffb4eb506dac5aba403f2b07c59002ebe1e8fc99ef5a843e15e17
3
+ size 15984
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c6815a247a59cc771fe631cb7e4fe6ea61917e06fbb1da84aad9abeb500fa27
3
+ size 1064
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 10.464989785409928,
3
+ "best_model_checkpoint": "./whisper-large-v3-praxy-fr-30k-teeeest-final/checkpoint-5",
4
+ "epoch": 0.14814814814814814,
5
+ "eval_steps": 5,
6
+ "global_step": 5,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.02962962962962963,
13
+ "grad_norm": 0.7531973123550415,
14
+ "learning_rate": 2e-08,
15
+ "loss": 0.0613,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.05925925925925926,
20
+ "grad_norm": 0.6711551547050476,
21
+ "learning_rate": 4e-08,
22
+ "loss": 0.0599,
23
+ "step": 2
24
+ },
25
+ {
26
+ "epoch": 0.08888888888888889,
27
+ "grad_norm": 0.8161910772323608,
28
+ "learning_rate": 6.000000000000001e-08,
29
+ "loss": 0.0514,
30
+ "step": 3
31
+ },
32
+ {
33
+ "epoch": 0.11851851851851852,
34
+ "grad_norm": 0.783392071723938,
35
+ "learning_rate": 8e-08,
36
+ "loss": 0.0656,
37
+ "step": 4
38
+ },
39
+ {
40
+ "epoch": 0.14814814814814814,
41
+ "grad_norm": 0.6716492176055908,
42
+ "learning_rate": 1.0000000000000001e-07,
43
+ "loss": 0.0505,
44
+ "step": 5
45
+ },
46
+ {
47
+ "epoch": 0.14814814814814814,
48
+ "eval_MediaSpeech_loss": 0.3272668123245239,
49
+ "eval_MediaSpeech_runtime": 65.3379,
50
+ "eval_MediaSpeech_samples_per_second": 1.913,
51
+ "eval_MediaSpeech_steps_per_second": 0.061,
52
+ "eval_MediaSpeech_wer": 13.706183536825522,
53
+ "step": 5
54
+ },
55
+ {
56
+ "epoch": 0.14814814814814814,
57
+ "eval_Pxcorpus_loss": 0.19093938171863556,
58
+ "eval_Pxcorpus_runtime": 30.7203,
59
+ "eval_Pxcorpus_samples_per_second": 1.302,
60
+ "eval_Pxcorpus_steps_per_second": 0.065,
61
+ "eval_Pxcorpus_wer": 7.223796033994335,
62
+ "step": 5
63
+ }
64
+ ],
65
+ "logging_steps": 1,
66
+ "max_steps": 5,
67
+ "num_input_tokens_seen": 0,
68
+ "num_train_epochs": 1,
69
+ "save_steps": 5,
70
+ "stateful_callbacks": {
71
+ "EarlyStoppingCallback": {
72
+ "args": {
73
+ "early_stopping_patience": 3,
74
+ "early_stopping_threshold": 0.01
75
+ },
76
+ "attributes": {
77
+ "early_stopping_patience_counter": 0
78
+ }
79
+ },
80
+ "TrainerControl": {
81
+ "args": {
82
+ "should_epoch_stop": false,
83
+ "should_evaluate": false,
84
+ "should_log": false,
85
+ "should_save": true,
86
+ "should_training_stop": true
87
+ },
88
+ "attributes": {}
89
+ }
90
+ },
91
+ "total_flos": 2.1743988336544973e+18,
92
+ "train_batch_size": 4,
93
+ "trial_name": null,
94
+ "trial_params": null
95
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef0ee7f56988a560d7d0b7d389284d63b441b798ea81d224e78ddaa18d325cee
3
+ size 5432