hi000000 commited on
Commit
a612071
1 Parent(s): e4d3812

Upload tokenizer

Browse files
Files changed (2) hide show
  1. special_tokens_map.json +7 -1
  2. tokenizer_config.json +1 -1
special_tokens_map.json CHANGED
@@ -20,7 +20,13 @@
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
- "pad_token": "</s>",
 
 
 
 
 
 
24
  "unk_token": {
25
  "content": "<unk>",
26
  "lstrip": false,
 
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
+ "pad_token": {
24
+ "content": "<pad>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
  "unk_token": {
31
  "content": "<unk>",
32
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -1838,7 +1838,7 @@
1838
  "eos_token": "</s>",
1839
  "mask_token": "<mask>",
1840
  "model_max_length": 1000000000000000019884624838656,
1841
- "pad_token": "</s>",
1842
  "tokenizer_class": "PreTrainedTokenizerFast",
1843
  "unk_token": "<unk>"
1844
  }
 
1838
  "eos_token": "</s>",
1839
  "mask_token": "<mask>",
1840
  "model_max_length": 1000000000000000019884624838656,
1841
+ "pad_token": "<pad>",
1842
  "tokenizer_class": "PreTrainedTokenizerFast",
1843
  "unk_token": "<unk>"
1844
  }