ghofrani commited on
Commit
7100c32
1 Parent(s): a88b9f6

Training in progress, step 1000

Browse files
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<s>": 123, "</s>": 124}
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-xls-r-300m",
3
  "activation_dropout": 0.1,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
 
1
  {
2
+ "_name_or_path": "common7/checkpoint-500",
3
  "activation_dropout": 0.1,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d24f2a8a64fa4d2c417acb26435d2d51196e671bbead4e04abeb6613392a7c3e
3
  size 1262436145
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:933d64bca20af5a09f6acbcacc8c8e7d7ccaf0a226b04493bc7f2cda69cfa5dc
3
  size 1262436145
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "common7/", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f835d9a51063daca768f19bae6325db78e51b961675f3a6bc32e1bae7309d15
3
  size 2991
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:127236561e5b823c7d28b0ce4d6d6bc5b123912004e703efc887c6508f33a65e
3
  size 2991
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_": 1, "a": 2, "b": 3, "c": 4, "d": 5, "e": 6, "f": 7, "g": 8, "h": 9, "i": 10, "k": 11, "l": 12, "m": 13, "n": 14, "o": 15, "p": 16, "q": 17, "r": 18, "s": 19, "t": 20, "u": 21, "v": 22, "w": 23, "x": 24, "y": 25, "z": 26, "«": 27, "»": 28, "ā": 29, "š": 30, "،": 31, "؛": 32, "؟": 33, "ء": 34, "آ": 35, "أ": 36, "ؤ": 37, "ئ": 38, "ا": 39, "ب": 40, "ة": 41, "ت": 42, "ث": 43, "ج": 44, "ح": 45, "خ": 46, "د": 47, "ذ": 48, "ر": 49, "ز": 50, "س": 51, "ش": 52, "ص": 53, "ض": 54, "ط": 55, "ظ": 56, "ع": 57, "غ": 58, "ـ": 59, "ف": 60, "ق": 61, "ك": 62, "ل": 63, "م": 64, "ن": 65, "ه": 66, "و": 67, "ى": 68, "ي": 69, "ً": 70, "ٌ": 71, "َ": 72, "ُ": 73, "ِ": 74, "ّ": 75, "ْ": 76, "ٔ": 77, "٬": 78, "پ": 79, "چ": 80, "ژ": 81, "ک": 82, "گ": 83, "ۀ": 84, "ی": 85, "ے": 86, "ە": 87, "ﭘ": 88, "ﮐ": 89, "ﮔ": 90, "ﯽ": 91, "ﯾ": 92, "ﯿ": 93, "ﺍ": 94, "ﺎ": 95, "ﺑ": 96, "ﺒ": 97, "ﺖ": 98, "ﺘ": 99, "ﺧ": 100, "ﺩ": 101, "ﺪ": 102, "ﺭ": 103, "ﺮ": 104, "ﺱ": 105, "ﺴ": 106, "ﺷ": 107, "ﺸ": 108, "ﻀ": 109, "ﻋ": 110, "ﻌ": 111, "ﻟ": 112, "ﻡ": 113, "ﻢ": 114, "ﻤ": 115, "ﻥ": 116, "ﻧ": 117, "ﻪ": 118, "ﻭ": 119, "ﻮ": 120, "|": 0, "[UNK]": 121, "[PAD]": 122}