Alexander Markov commited on
Commit
c9144de
1 Parent(s): da4dc8e

updated model

Browse files
Files changed (4) hide show
  1. config.json +1 -1
  2. pytorch_model.bin +2 -2
  3. tokenizer.json +13 -13
  4. tokenizer_config.json +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./context-russian-inappropriate-messages/checkpoint-584/",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "context-russian-inappropriate-messages/",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f83b623726d48538a077857238102133c33838562cb1b7706c07551ab291392
3
- size 653868081
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13065d0cededdc25f4253ef9a14856dd8c3a8958dec4fc13ca01ba9e3c0a52c6
3
+ size 653881965
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 127,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
@@ -10,57 +10,57 @@
10
  "added_tokens": [
11
  {
12
  "id": 0,
 
13
  "content": "[PAD]",
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
- "special": true
19
  },
20
  {
21
  "id": 100,
 
22
  "content": "[UNK]",
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
- "special": true
28
  },
29
  {
30
  "id": 101,
 
31
  "content": "[CLS]",
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
- "special": true
37
  },
38
  {
39
  "id": 102,
 
40
  "content": "[SEP]",
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
- "special": true
46
  },
47
  {
48
  "id": 103,
 
49
  "content": "[MASK]",
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
  },
56
  {
57
  "id": 100792,
 
58
  "content": "[RESPONSE_TOKEN]",
59
  "single_word": false,
60
  "lstrip": false,
61
  "rstrip": false,
62
- "normalized": false,
63
- "special": true
64
  }
65
  ],
66
  "normalizer": {
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 128,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
10
  "added_tokens": [
11
  {
12
  "id": 0,
13
+ "special": true,
14
  "content": "[PAD]",
15
  "single_word": false,
16
  "lstrip": false,
17
  "rstrip": false,
18
+ "normalized": false
 
19
  },
20
  {
21
  "id": 100,
22
+ "special": true,
23
  "content": "[UNK]",
24
  "single_word": false,
25
  "lstrip": false,
26
  "rstrip": false,
27
+ "normalized": false
 
28
  },
29
  {
30
  "id": 101,
31
+ "special": true,
32
  "content": "[CLS]",
33
  "single_word": false,
34
  "lstrip": false,
35
  "rstrip": false,
36
+ "normalized": false
 
37
  },
38
  {
39
  "id": 102,
40
+ "special": true,
41
  "content": "[SEP]",
42
  "single_word": false,
43
  "lstrip": false,
44
  "rstrip": false,
45
+ "normalized": false
 
46
  },
47
  {
48
  "id": 103,
49
+ "special": true,
50
  "content": "[MASK]",
51
  "single_word": false,
52
  "lstrip": false,
53
  "rstrip": false,
54
+ "normalized": false
 
55
  },
56
  {
57
  "id": 100792,
58
+ "special": true,
59
  "content": "[RESPONSE_TOKEN]",
60
  "single_word": false,
61
  "lstrip": false,
62
  "rstrip": false,
63
+ "normalized": false
 
64
  }
65
  ],
66
  "normalizer": {
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "do_basic_tokenize": true, "never_split": null, "use_fast": true, "special_tokens_map_file": "/root/.cache/huggingface/transformers/1f428acdde727eed5de979d6856ce350a470be2a64e134a1fdae04af78a27301.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "./context-russian-inappropriate-messages/checkpoint-584/", "tokenizer_class": "BertTokenizer"}
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "do_basic_tokenize": true, "never_split": null, "use_fast": true, "special_tokens_map_file": "/root/.cache/huggingface/transformers/1f428acdde727eed5de979d6856ce350a470be2a64e134a1fdae04af78a27301.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "context-russian-inappropriate-messages/", "tokenizer_class": "BertTokenizer"}