patrickvonplaten commited on
Commit
ce4eb45
1 Parent(s): 7872b1a

update tokenizer

Browse files
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<s>": 8065, "</s>": 8066}
feature_extractor_config.json DELETED
@@ -1,13 +0,0 @@
1
- {
2
- "do_normalize": true,
3
- "feature_size": 80,
4
- "sampling_rate": 16000,
5
- "padding_value": 0.0,
6
- "hop_length": 10,
7
- "win_length": 25,
8
- "win_function": "hamming_window",
9
- "normalize_means": true,
10
- "normalize_vars": true,
11
- "return_attention_mask": false
12
- }
13
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
preprocessor_config.json CHANGED
@@ -1,12 +1,21 @@
1
  {
 
2
  "do_normalize": true,
 
3
  "feature_size": 80,
4
- "sampling_rate": 16000,
5
- "padding_value": 0.0,
6
  "hop_length": 10,
7
- "win_length": 25,
8
- "win_function": "hamming_window",
9
  "normalize_means": true,
10
- "normalize_vars": true,
11
- "return_attention_mask": false
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "K": 257,
3
  "do_normalize": true,
4
+ "feature_extractor_type": "MCTCFeatureExtractor",
5
  "feature_size": 80,
6
+ "frame_signal_scale": 32768.0,
 
7
  "hop_length": 10,
8
+ "mel_floor": 1.0,
9
+ "n_fft": 512,
10
  "normalize_means": true,
11
+ "normalize_vars": true,
12
+ "padding_side": "right",
13
+ "padding_value": 0.0,
14
+ "preemphasis_coeff": 0.97,
15
+ "return_attention_mask": false,
16
+ "sample_size": 400,
17
+ "sample_stride": 160,
18
+ "sampling_rate": 16000,
19
+ "win_function": "hamming_window",
20
+ "win_length": 25
21
  }
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "return_attention_mask": false, "do_normalize": true}
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "word_delimiter_token": "|", "replace_word_delimiter_char": " ", "return_attention_mask": false, "do_normalize": true, "special_tokens_map_file": "./mctc-large/special_tokens_map.json", "name_or_path": "./mctc-large", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
vocab.json CHANGED
The diff for this file is too large to render. See raw diff