Training in progress, step 500
Browse files- .ipynb_checkpoints/preprocessor_config-checkpoint.json +9 -0
- .ipynb_checkpoints/train_results-checkpoint.json +8 -0
- .ipynb_checkpoints/vocab-checkpoint.json +1 -0
- added_tokens.json +1 -0
- config.json +2 -2
- pytorch_model.bin +2 -2
- runs/Jan28_17-40-32_job-0074bb36-c67f-4775-b1b6-176eb09b0ba4/1643391738.2800581/events.out.tfevents.1643391738.job-0074bb36-c67f-4775-b1b6-176eb09b0ba4.843454.1 +3 -0
- runs/Jan28_17-40-32_job-0074bb36-c67f-4775-b1b6-176eb09b0ba4/events.out.tfevents.1643391738.job-0074bb36-c67f-4775-b1b6-176eb09b0ba4.843454.0 +3 -0
- special_tokens_map.json +1 -0
- tokenizer_config.json +1 -0
- training_args.bin +1 -1
- vocab.json +1 -0
.ipynb_checkpoints/preprocessor_config-checkpoint.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"do_normalize": true,
|
3 |
+
"feature_extractor_type": "Wav2Vec2FeatureExtractor",
|
4 |
+
"feature_size": 1,
|
5 |
+
"padding_side": "right",
|
6 |
+
"padding_value": 0,
|
7 |
+
"return_attention_mask": true,
|
8 |
+
"sampling_rate": 16000
|
9 |
+
}
|
.ipynb_checkpoints/train_results-checkpoint.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 50.0,
|
3 |
+
"train_loss": 1.5738848333155855,
|
4 |
+
"train_runtime": 24806.553,
|
5 |
+
"train_samples": 6034,
|
6 |
+
"train_samples_per_second": 12.162,
|
7 |
+
"train_steps_per_second": 0.379
|
8 |
+
}
|
.ipynb_checkpoints/vocab-checkpoint.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"،": 1, "؛": 2, "؟": 3, "ئ": 4, "ا": 5, "ب": 6, "ت": 7, "ج": 8, "خ": 9, "د": 10, "ر": 11, "ز": 12, "س": 13, "ش": 14, "غ": 15, "ف": 16, "ق": 17, "ك": 18, "ل": 19, "م": 20, "ن": 21, "و": 22, "ى": 23, "ي": 24, "پ": 25, "چ": 26, "ژ": 27, "ڭ": 28, "گ": 29, "ھ": 30, "ۆ": 31, "ۇ": 32, "ۈ": 33, "ۋ": 34, "ې": 35, "ە": 36, "‹": 37, "›": 38, "−": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
|
added_tokens.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"<s>": 42, "</s>": 43}
|
config.json
CHANGED
@@ -76,7 +76,7 @@
|
|
76 |
"num_hidden_layers": 24,
|
77 |
"num_negatives": 100,
|
78 |
"output_hidden_size": 1024,
|
79 |
-
"pad_token_id":
|
80 |
"proj_codevector_dim": 768,
|
81 |
"tdnn_dilation": [
|
82 |
1,
|
@@ -102,6 +102,6 @@
|
|
102 |
"torch_dtype": "float32",
|
103 |
"transformers_version": "4.16.0.dev0",
|
104 |
"use_weighted_layer_sum": false,
|
105 |
-
"vocab_size":
|
106 |
"xvector_output_dim": 512
|
107 |
}
|
|
|
76 |
"num_hidden_layers": 24,
|
77 |
"num_negatives": 100,
|
78 |
"output_hidden_size": 1024,
|
79 |
+
"pad_token_id": 41,
|
80 |
"proj_codevector_dim": 768,
|
81 |
"tdnn_dilation": [
|
82 |
1,
|
|
|
102 |
"torch_dtype": "float32",
|
103 |
"transformers_version": "4.16.0.dev0",
|
104 |
"use_weighted_layer_sum": false,
|
105 |
+
"vocab_size": 44,
|
106 |
"xvector_output_dim": 512
|
107 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:229d18875aa5551d40c74956fa68d813fbb0cdcf4e488137e895d56fefa1cae6
|
3 |
+
size 1262104049
|
runs/Jan28_17-40-32_job-0074bb36-c67f-4775-b1b6-176eb09b0ba4/1643391738.2800581/events.out.tfevents.1643391738.job-0074bb36-c67f-4775-b1b6-176eb09b0ba4.843454.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b097f29a6bc269d898423d71c69267f61a1dfe37e41799508bb02ded928bd248
|
3 |
+
size 4802
|
runs/Jan28_17-40-32_job-0074bb36-c67f-4775-b1b6-176eb09b0ba4/events.out.tfevents.1643391738.job-0074bb36-c67f-4775-b1b6-176eb09b0ba4.843454.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6fbd8a92cb461a252a5c5c12c6026b6848e17b94dd2ed6e80d159342e221a5a
|
3 |
+
size 5830
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./xls-r-uyghur-cv8", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3055
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81c852d75cc7308b223dd2f8668f73ae311d16d0fb416266206ec9a8c702837a
|
3 |
size 3055
|
vocab.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"،": 1, "؛": 2, "؟": 3, "ئ": 4, "ا": 5, "ب": 6, "ت": 7, "ج": 8, "خ": 9, "د": 10, "ر": 11, "ز": 12, "س": 13, "ش": 14, "غ": 15, "ف": 16, "ق": 17, "ك": 18, "ل": 19, "م": 20, "ن": 21, "و": 22, "ى": 23, "ي": 24, "پ": 25, "چ": 26, "ژ": 27, "ڭ": 28, "گ": 29, "ھ": 30, "ۆ": 31, "ۇ": 32, "ۈ": 33, "ۋ": 34, "ې": 35, "ە": 36, "‹": 37, "›": 38, "−": 39, "|": 0, "[UNK]": 40, "[PAD]": 41}
|