davidmasip commited on
Commit
c7894d3
1 Parent(s): 0fde064

long training

Browse files
config.json CHANGED
@@ -3,18 +3,18 @@
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
- "hidden_dropout_prob": 0.1,
13
  "hidden_size": 768,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
16
- "layer_norm_eps": 1e-12,
17
- "max_position_embeddings": 130,
18
  "model_type": "roberta",
19
  "num_attention_heads": 12,
20
  "num_hidden_layers": 12,
@@ -25,5 +25,5 @@
25
  "transformers_version": "4.16.2",
26
  "type_vocab_size": 1,
27
  "use_cache": true,
28
- "vocab_size": 30002
29
  }
 
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.0,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.0,
13
  "hidden_size": 768,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
  "model_type": "roberta",
19
  "num_attention_heads": 12,
20
  "num_hidden_layers": 12,
 
25
  "transformers_version": "4.16.2",
26
  "type_vocab_size": 1,
27
  "use_cache": true,
28
+ "vocab_size": 50262
29
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0612630ffcfec52f6df89d77f7da345dfb7b463a8b2c1a31555184972def7b25
3
- size 435240621
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d1510ecd75b2a89fae8d6bc7d5a320c5385f0b6084dba1292893226a7f5126f
3
+ size 498662061
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": "<mask>"}
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "special_tokens_map_file": "models/twerto-base-uncased/special_tokens_map.json", "name_or_path": "racism", "tokenizer_class": "PreTrainedTokenizerFast"}
 
1
+ {"unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "errors": "replace", "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "trim_offsets": true, "max_len": 512, "special_tokens_map_file": null, "name_or_path": "racism", "tokenizer_class": "RobertaTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92f8a532727615ff8d4d694d12984a7d34af19c4172dd2739eada23b8787cf21
3
  size 2991
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:606d24b6cde093dd2923e73f4873793131ed09c810f9bd02dbb39a85bfc66664
3
  size 2991
vocab.json ADDED
The diff for this file is too large to render. See raw diff