D1V1DE commited on
Commit
d19eb04
·
verified ·
1 Parent(s): cd28198

Delete bias-detection-tokenizer

Browse files
bias-detection-tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
bias-detection-tokenizer/special_tokens_map.json DELETED
@@ -1,15 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bias-detection-tokenizer/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
bias-detection-tokenizer/tokenizer_config.json DELETED
@@ -1,16 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": "<s>",
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "errors": "replace",
7
- "mask_token": "<mask>",
8
- "model_max_length": 512,
9
- "name_or_path": "valurank/distilroberta-bias",
10
- "pad_token": "<pad>",
11
- "sep_token": "</s>",
12
- "special_tokens_map_file": null,
13
- "tokenizer_class": "RobertaTokenizer",
14
- "trim_offsets": true,
15
- "unk_token": "<unk>"
16
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bias-detection-tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff