jesseD commited on
Commit
44b5694
1 Parent(s): cba3463

add tokenizer

Browse files
Files changed (2) hide show
  1. tokenizer.json +4 -4
  2. tokenizer_config.json +1 -1
tokenizer.json CHANGED
@@ -19,21 +19,21 @@
19
  "added_tokens": [
20
  {
21
  "id": 50256,
22
- "special": true,
23
  "content": "<|endoftext|>",
24
  "single_word": false,
25
  "lstrip": false,
26
  "rstrip": false,
27
- "normalized": true
 
28
  },
29
  {
30
  "id": 50257,
31
- "special": true,
32
  "content": "<|pad|>",
33
  "single_word": false,
34
  "lstrip": false,
35
  "rstrip": false,
36
- "normalized": false
 
37
  }
38
  ],
39
  "normalizer": null,
19
  "added_tokens": [
20
  {
21
  "id": 50256,
 
22
  "content": "<|endoftext|>",
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": true,
27
+ "special": true
28
  },
29
  {
30
  "id": 50257,
 
31
  "content": "<|pad|>",
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": false,
36
+ "special": true
37
  }
38
  ],
39
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "return_special_tokens_mask": true, "special_tokens_map_file": null, "name_or_path": "microsoft/DialoGPT-medium", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}
1
+ {"unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "model_max_length": 1024, "return_special_tokens_mask": true, "special_tokens_map_file": null, "name_or_path": "DingleyMaillotUrgell/homer-bot", "errors": "replace", "tokenizer_class": "GPT2Tokenizer"}