peter2000 commited on
Commit
5eb8cfc
1 Parent(s): 74911cc

Training in progress, step 400

Browse files
config.json CHANGED
@@ -5,7 +5,7 @@
5
  "architectures": [
6
  "Wav2Vec2ForCTC"
7
  ],
8
- "attention_dropout": 0.0,
9
  "bos_token_id": 1,
10
  "classifier_proj_size": 256,
11
  "codevector_dim": 768,
@@ -46,21 +46,21 @@
46
  "feat_extract_activation": "gelu",
47
  "feat_extract_dropout": 0.0,
48
  "feat_extract_norm": "layer",
49
- "feat_proj_dropout": 0.0,
50
  "feat_quantizer_dropout": 0.0,
51
  "final_dropout": 0.0,
52
  "gradient_checkpointing": false,
53
  "hidden_act": "gelu",
54
- "hidden_dropout": 0.0,
55
  "hidden_size": 1024,
56
  "initializer_range": 0.02,
57
  "intermediate_size": 4096,
58
  "layer_norm_eps": 1e-05,
59
- "layerdrop": 0.0,
60
  "mask_feature_length": 10,
61
  "mask_feature_prob": 0.0,
62
  "mask_time_length": 10,
63
- "mask_time_prob": 0.0,
64
  "model_type": "wav2vec2",
65
  "num_attention_heads": 16,
66
  "num_codevector_groups": 2,
@@ -70,7 +70,7 @@
70
  "num_feat_extract_layers": 7,
71
  "num_hidden_layers": 24,
72
  "num_negatives": 100,
73
- "pad_token_id": 97,
74
  "proj_codevector_dim": 768,
75
  "torch_dtype": "float32",
76
  "transformers_version": "4.11.3",
 
5
  "architectures": [
6
  "Wav2Vec2ForCTC"
7
  ],
8
+ "attention_dropout": 0.05,
9
  "bos_token_id": 1,
10
  "classifier_proj_size": 256,
11
  "codevector_dim": 768,
 
46
  "feat_extract_activation": "gelu",
47
  "feat_extract_dropout": 0.0,
48
  "feat_extract_norm": "layer",
49
+ "feat_proj_dropout": 0.05,
50
  "feat_quantizer_dropout": 0.0,
51
  "final_dropout": 0.0,
52
  "gradient_checkpointing": false,
53
  "hidden_act": "gelu",
54
+ "hidden_dropout": 0.05,
55
  "hidden_size": 1024,
56
  "initializer_range": 0.02,
57
  "intermediate_size": 4096,
58
  "layer_norm_eps": 1e-05,
59
+ "layerdrop": 0.05,
60
  "mask_feature_length": 10,
61
  "mask_feature_prob": 0.0,
62
  "mask_time_length": 10,
63
+ "mask_time_prob": 0.05,
64
  "model_type": "wav2vec2",
65
  "num_attention_heads": 16,
66
  "num_codevector_groups": 2,
 
70
  "num_feat_extract_layers": 7,
71
  "num_hidden_layers": 24,
72
  "num_negatives": 100,
73
+ "pad_token_id": 96,
74
  "proj_codevector_dim": 768,
75
  "torch_dtype": "float32",
76
  "transformers_version": "4.11.3",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad8e5f585dd9d5cde0c6bdd6c0a5eb5e59b2fa60227233428ea7d6170445ceac
3
  size 1262329585
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e21329aad1c5fed11709ef5195f1a219643294ca021346b1037c441823463837
3
  size 1262329585
runs/Jun02_19-30-03_e97f4e46d604/1654200508.2227576/events.out.tfevents.1654200508.e97f4e46d604.97.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:851365652599db95f0b4982551ed0eb33d6afc9d31aa334eee72781cf5e64753
3
+ size 4606
runs/Jun02_19-30-03_e97f4e46d604/events.out.tfevents.1654200508.e97f4e46d604.97.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44da1dd19a2bf2051ba40b37368195daac1a5bd346ef73b658c05c662aad8fa1
3
+ size 4738
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42fdf4eca0a874456388456fbb4426785f36ea4a0d879bd1ac39266c6e8b8647
3
  size 2863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a796dd1f339098f7a8f13dc3b230033f2f9173516215606d175c9af95d8bcd5
3
  size 2863