Konrad Wojtasik commited on
Commit
d904572
1 Parent(s): aaecec7
CERerankingEvaluator_train-eval_results.csv ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MRR@10
2
+ 0,10000,0.5516428571428571
3
+ 0,20000,0.5693511904761904
4
+ 0,30000,0.5603849206349206
5
+ 0,40000,0.5249920634920635
6
+ 0,50000,0.501452380952381
7
+ 0,60000,0.5223412698412698
8
+ 0,70000,0.019930555555555552
9
+ 0,80000,0.1
10
+ 0,90000,0.5030555555555555
11
+ 0,100000,0.985
12
+ 0,110000,0.24
13
+ 0,120000,0.995
14
+ 0,130000,0.135
15
+ 0,140000,0.06330555555555556
16
+ 0,150000,0.5398055555555555
17
+ 0,160000,0.785
18
+ 0,170000,0.3506666666666666
19
+ 0,180000,0.1963452380952381
20
+ 0,190000,0.08798611111111111
21
+ 0,200000,0.8279166666666665
22
+ 0,210000,0.4415833333333333
23
+ 0,220000,1.0
24
+ 0,230000,0.3698095238095238
25
+ 0,240000,0.43
26
+ 0,250000,0.35
27
+ 0,260000,0.06287698412698413
28
+ 0,270000,0.99
29
+ 0,280000,0.885
30
+ 0,290000,1.0
31
+ 0,300000,0.945
32
+ 0,310000,0.985
33
+ 0,320000,0.7991666666666666
34
+ 0,330000,1.0
35
+ 0,340000,0.43
36
+ 0,350000,0.175
37
+ 0,360000,0.9225
38
+ 0,370000,1.0
39
+ 0,380000,0.575
40
+ 0,390000,1.0
41
+ 0,400000,0.205
42
+ 0,410000,1.0
43
+ 0,420000,0.5507500000000001
44
+ 0,430000,1.0
45
+ 0,440000,1.0
46
+ 0,450000,1.0
47
+ 0,460000,0.41
48
+ 0,470000,0.995
49
+ 0,480000,1.0
50
+ 0,490000,0.06780158730158731
51
+ 0,500000,0.4
52
+ 0,510000,0.77
53
+ 0,520000,1.0
54
+ 0,530000,0.8276666666666666
55
+ 0,540000,1.0
56
+ 0,550000,0.5896666666666667
57
+ 0,560000,0.9925
58
+ 0,570000,0.45
59
+ 0,580000,0.8
60
+ 0,590000,1.0
61
+ 0,600000,0.21767063492063493
62
+ 0,610000,0.9591666666666667
63
+ 0,620000,1.0
64
+ 0,-1,0.995
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "allegro/herbert-large-cased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 4096,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 514,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 16,
24
+ "num_hidden_layers": 24,
25
+ "pad_token_id": 1,
26
+ "pooler_fc_size": 768,
27
+ "pooler_num_attention_heads": 12,
28
+ "pooler_num_fc_layers": 3,
29
+ "pooler_size_per_head": 128,
30
+ "pooler_type": "first_token_transform",
31
+ "position_embedding_type": "absolute",
32
+ "tokenizer_class": "HerbertTokenizerFast",
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.26.1",
35
+ "type_vocab_size": 2,
36
+ "use_cache": true,
37
+ "vocab_size": 50000
38
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3e7d887dcbc975078374c8e4c3d8131b9e7ae63f21330096c111a784989c0dd
3
+ size 1420500597
special_tokens_map.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "mask_token": "<mask>",
5
+ "pad_token": "<pad>",
6
+ "sep_token": "</s>",
7
+ "unk_token": "<unk>"
8
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [],
3
+ "bos_token": "<s>",
4
+ "cls_token": "<s>",
5
+ "do_lowercase_and_remove_accent": false,
6
+ "id2lang": null,
7
+ "lang2id": null,
8
+ "mask_token": "<mask>",
9
+ "model_max_length": 512,
10
+ "name_or_path": "allegro/herbert-large-cased",
11
+ "pad_token": "<pad>",
12
+ "sep_token": "</s>",
13
+ "special_tokens_map_file": "/root/.cache/huggingface/hub/models--allegro--herbert-large-cased/snapshots/8d0fa3bc0566c3a332bec0d471c8d8c37b5cbb90/special_tokens_map.json",
14
+ "tokenizer_class": "HerbertTokenizer",
15
+ "unk_token": "<unk>"
16
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff