nishantyadav commited on
Commit
6226c92
1 Parent(s): 2ba776b

Upload 8 files

Browse files

Adding model files

CERerankingEvaluator_train-eval_results.csv ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MRR@10
2
+ 0,5000,0.5878849206349206
3
+ 0,10000,0.6041547619047618
4
+ 0,15000,0.6497738095238096
5
+ 0,20000,0.6110515873015874
6
+ 0,25000,0.6367162698412698
7
+ 0,30000,0.6215198412698413
8
+ 0,35000,0.6465575396825397
9
+ 0,40000,0.6287678571428572
10
+ 0,45000,0.6327579365079364
11
+ 0,50000,0.6261111111111112
12
+ 0,55000,0.6344345238095238
13
+ 0,60000,0.6084484126984127
14
+ 0,65000,0.6272738095238095
15
+ 0,70000,0.6481785714285715
16
+ 0,75000,0.644126984126984
17
+ 0,80000,0.6228531746031746
18
+ 0,85000,0.6537281746031745
19
+ 0,90000,0.6524146825396826
20
+ 0,95000,0.6453988095238095
21
+ 0,100000,0.6594503968253967
22
+ 0,105000,0.6702956349206349
23
+ 0,110000,0.6492400793650793
24
+ 0,115000,0.67440873015873
25
+ 0,120000,0.6352738095238095
26
+ 0,125000,0.6509444444444444
27
+ 0,130000,0.6605456349206349
28
+ 0,135000,0.6842757936507937
29
+ 0,140000,0.6641765873015874
30
+ 0,145000,0.6676964285714285
31
+ 0,150000,0.6606488095238094
32
+ 0,155000,0.6590317460317461
33
+ 0,160000,0.654829365079365
34
+ 0,165000,0.6451646825396824
35
+ 0,170000,0.6600734126984128
36
+ 0,175000,0.6531686507936508
37
+ 0,180000,0.630827380952381
38
+ 0,185000,0.6735079365079365
39
+ 0,190000,0.651704365079365
40
+ 0,195000,0.662265873015873
41
+ 0,200000,0.6383511904761905
42
+ 0,205000,0.6420714285714285
43
+ 0,210000,0.6305972222222223
44
+ 0,215000,0.6452301587301587
45
+ 0,220000,0.6744999999999999
46
+ 0,225000,0.6822460317460317
47
+ 0,230000,0.6566746031746032
48
+ 0,235000,0.6805257936507937
49
+ 0,240000,0.6503313492063492
50
+ 0,245000,0.6600714285714285
51
+ 0,250000,0.6600019841269841
52
+ 0,255000,0.6622936507936508
53
+ 0,260000,0.5985892857142857
54
+ 0,265000,0.6323353174603175
55
+ 0,270000,0.6193690476190475
56
+ 0,275000,0.6655218253968255
57
+ 0,280000,0.6651488095238095
58
+ 0,285000,0.6695575396825396
59
+ 0,290000,0.6623710317460317
60
+ 0,295000,0.6863551587301586
61
+ 0,300000,0.6858849206349206
62
+ 0,305000,0.6696626984126983
63
+ 0,310000,0.6591210317460318
64
+ 0,315000,0.6722123015873015
65
+ 0,320000,0.6617321428571428
66
+ 0,325000,0.6423472222222222
67
+ 0,330000,0.6636547619047619
68
+ 0,335000,0.6625853174603173
69
+ 0,340000,0.6576289682539683
70
+ 0,345000,0.6697599206349207
71
+ 0,350000,0.6718075396825395
72
+ 0,355000,0.6753373015873015
73
+ 0,360000,0.6596825396825398
74
+ 0,365000,0.6624007936507936
75
+ 0,370000,0.6530932539682539
76
+ 0,375000,0.6472222222222223
77
+ 0,380000,0.6605357142857141
78
+ 0,385000,0.6657361111111111
79
+ 0,390000,0.6669206349206348
80
+ 0,395000,0.6542281746031746
81
+ 0,400000,0.6495952380952381
82
+ 0,405000,0.6605694444444444
83
+ 0,410000,0.6495734126984126
84
+ 0,415000,0.6558154761904763
85
+ 0,420000,0.6588869047619048
86
+ 0,425000,0.6399623015873016
87
+ 0,430000,0.6456706349206348
88
+ 0,435000,0.6531845238095237
89
+ 0,440000,0.6456706349206348
90
+ 0,445000,0.6548789682539683
91
+ 0,450000,0.6497738095238094
92
+ 0,455000,0.6543769841269841
93
+ 0,460000,0.6469126984126984
94
+ 0,465000,0.656593253968254
95
+ 0,470000,0.6423630952380952
96
+ 0,475000,0.6589246031746032
97
+ 0,480000,0.655968253968254
98
+ 0,485000,0.6469246031746032
99
+ 0,490000,0.6506488095238097
100
+ 0,495000,0.6581746031746032
101
+ 0,500000,0.667265873015873
102
+ 0,505000,0.6603214285714284
103
+ 0,510000,0.6564623015873017
104
+ 0,515000,0.6651765873015872
105
+ 0,520000,0.6671726190476189
106
+ 0,525000,0.6602222222222222
107
+ 0,530000,0.6602440476190476
108
+ 0,535000,0.6694603174603173
109
+ 0,540000,0.6527797619047618
110
+ 0,545000,0.6576130952380953
111
+ 0,550000,0.6624821428571428
112
+ 0,555000,0.6621825396825396
113
+ 0,560000,0.6427718253968254
114
+ 0,565000,0.6513571428571429
115
+ 0,570000,0.6510575396825397
116
+ 0,575000,0.6631071428571428
117
+ 0,580000,0.6439821428571428
118
+ 0,585000,0.6595793650793651
119
+ 0,590000,0.6606011904761906
120
+ 0,595000,0.652126984126984
121
+ 0,600000,0.6547460317460317
122
+ 0,605000,0.6541626984126984
123
+ 0,610000,0.650611111111111
124
+ 0,615000,0.6470634920634921
125
+ 0,620000,0.6496329365079364
126
+ 0,625000,0.6470634920634921
127
+ 0,-1,0.6470634920634921
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "bert",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "output_hidden_states": true,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.26.1",
30
+ "type_vocab_size": 2,
31
+ "use_cache": true,
32
+ "vocab_size": 30522
33
+ }
orig_param_for_run.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name": "bert-base-uncased",
3
+ "use_embed_ce_model": 1,
4
+ "res_dir": "../../results/9_BEIR_CrossEnc/basemodel=bert_base_ce_pool=embed_loss=bce_from_scratch_seed=0",
5
+ "evaluation_steps": 5000,
6
+ "train_batch_size": 32,
7
+ "lr": 2e-05,
8
+ "disable_wandb": 0,
9
+ "seed": 0
10
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b71a4c8448503649a26c38e7ae273df1d699d6a8504a0a629ae60d0a817502b6
3
+ size 438003053
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff