add tokenizer
Browse files- tokenizer_config.json +1 -1
- vocab.json +1 -1
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
|
|
1 |
+
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "replace_word_delimiter_char": " ", "tokenizer_class": "Wav2Vec2CTCTokenizer", "processor_class": "Wav2Vec2Processor"}
|
vocab.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"
|
|
|
1 |
+
{"0": 0, "n": 1, "4": 2, "z": 3, "7": 4, "m": 5, "r": 6, "(": 7, "c": 8, "2": 9, "a": 10, "j": 11, "o": 12, "y": 13, "g": 14, "q": 16, ")": 17, "e": 18, "i": 19, "5": 20, "v": 21, "s": 22, "=": 23, "b": 24, "f": 25, "w": 26, "6": 27, "k": 28, "h": 29, "d": 30, "p": 31, "x": 32, "/": 33, "1": 34, "u": 35, "l": 36, "t": 37, "|": 15, "[UNK]": 38, "[PAD]": 39}
|