dataautogpt3 commited on
Commit
5c35c52
1 Parent(s): 8abf599

Delete tokenizer

Browse files
tokenizer/tokenizer_added_tokens.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "<|endoftext|>": 49407,
3
- "<|startoftext|>": 49406
4
- }
 
 
 
 
 
tokenizer/tokenizer_config.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "added_tokens_decoder": {
4
- "49406": {
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": false
11
- },
12
- "49407": {
13
- "content": "<|endoftext|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": false
19
- }
20
- },
21
- "additional_special_tokens": [],
22
- "bos_token": "<|startoftext|>",
23
- "clean_up_tokenization_spaces": true,
24
- "do_lower_case": true,
25
- "eos_token": "<|endoftext|>",
26
- "errors": "replace",
27
- "model_max_length": 77,
28
- "pad_token": "<|endoftext|>",
29
- "tokenizer_class": "CLIPTokenizer",
30
- "tokenizer_file": "/home/user/.cache/huggingface/hub/models--openai--clip-vit-large-patch14/snapshots/32bd64288804d66eefd0ccbe215aa642df71cc41/tokenizer.json",
31
- "unk_token": "<|endoftext|>"
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/tokenizer_merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer/tokenizer_special_tokens_map.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token": "<|startoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
6
- }
 
 
 
 
 
 
 
tokenizer/tokenizer_vocab.json DELETED
The diff for this file is too large to render. See raw diff