saattrupdan
commited on
Commit
•
e860ee0
1
Parent(s):
67f752e
Training in progress, step 300
Browse files- config.json +4 -4
- pytorch_model.bin +2 -2
- tokenizer_config.json +1 -1
- training_args.bin +1 -1
- vocab.json +1 -1
config.json
CHANGED
@@ -9,7 +9,7 @@
|
|
9 |
"Wav2Vec2ForCTC"
|
10 |
],
|
11 |
"attention_dropout": 0.1,
|
12 |
-
"bos_token_id":
|
13 |
"classifier_proj_size": 256,
|
14 |
"codevector_dim": 768,
|
15 |
"contrastive_logits_temperature": 0.1,
|
@@ -45,7 +45,7 @@
|
|
45 |
"ctc_zero_infinity": false,
|
46 |
"diversity_loss_weight": 0.1,
|
47 |
"do_stable_layer_norm": true,
|
48 |
-
"eos_token_id":
|
49 |
"feat_extract_activation": "gelu",
|
50 |
"feat_extract_dropout": 0.0,
|
51 |
"feat_extract_norm": "layer",
|
@@ -77,7 +77,7 @@
|
|
77 |
"num_hidden_layers": 24,
|
78 |
"num_negatives": 100,
|
79 |
"output_hidden_size": 1024,
|
80 |
-
"pad_token_id":
|
81 |
"proj_codevector_dim": 768,
|
82 |
"tdnn_dilation": [
|
83 |
1,
|
@@ -103,6 +103,6 @@
|
|
103 |
"torch_dtype": "float32",
|
104 |
"transformers_version": "4.16.2",
|
105 |
"use_weighted_layer_sum": false,
|
106 |
-
"vocab_size":
|
107 |
"xvector_output_dim": 512
|
108 |
}
|
9 |
"Wav2Vec2ForCTC"
|
10 |
],
|
11 |
"attention_dropout": 0.1,
|
12 |
+
"bos_token_id": 33,
|
13 |
"classifier_proj_size": 256,
|
14 |
"codevector_dim": 768,
|
15 |
"contrastive_logits_temperature": 0.1,
|
45 |
"ctc_zero_infinity": false,
|
46 |
"diversity_loss_weight": 0.1,
|
47 |
"do_stable_layer_norm": true,
|
48 |
+
"eos_token_id": 34,
|
49 |
"feat_extract_activation": "gelu",
|
50 |
"feat_extract_dropout": 0.0,
|
51 |
"feat_extract_norm": "layer",
|
77 |
"num_hidden_layers": 24,
|
78 |
"num_negatives": 100,
|
79 |
"output_hidden_size": 1024,
|
80 |
+
"pad_token_id": 32,
|
81 |
"proj_codevector_dim": 768,
|
82 |
"tdnn_dilation": [
|
83 |
1,
|
103 |
"torch_dtype": "float32",
|
104 |
"transformers_version": "4.16.2",
|
105 |
"use_weighted_layer_sum": false,
|
106 |
+
"vocab_size": 35,
|
107 |
"xvector_output_dim": 512
|
108 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:faf1b89d9dd7fb9d3776b706fbbf062ef8a76771db60bc19154a0af211b90726
|
3 |
+
size 1262073098
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case":
|
1 |
+
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./", "tokenizer_class": "Wav2Vec2CTCTokenizer", "processor_class": "Wav2Vec2Processor"}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3055
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bac45e9b0784ba02fb32604ffb531b6713f8d2c6f19f68e4fa37fdfc89352595
|
3 |
size 3055
|
vocab.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"
|
1 |
+
{"h": 0, "r": 1, "l": 2, "u": 3, "a": 4, "v": 5, "|": 6, "ø": 7, "d": 8, "f": 9, "c": 10, "z": 11, "o": 12, "é": 13, "p": 14, "s": 15, "x": 16, "i": 17, "n": 18, "ó": 19, "b": 20, "e": 21, "w": 22, "k": 23, "æ": 24, "g": 25, "j": 26, "y": 27, "t": 28, "å": 29, "m": 30, "<unk>": 31, "<pad>": 32, "<s>": 33, "</s>": 34}
|