nreimers commited on
Commit
d7a7bc1
1 Parent(s): e60ab70
CEBinaryClassificationEvaluator_Quora-dev_results.csv ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,Accuracy,Accuracy_Threshold,F1,F1_Threshold,Precision,Recall,Average_Precision
2
+ 0,5000,0.8683833274728842,0.5240041017532349,0.8198757763975154,0.3768240511417389,0.7744990392533626,0.8709005324484914,0.8731992606250942
3
+ 0,10000,0.8685997132887939,0.6672844886779785,0.8178615982974976,0.4671083688735962,0.7796432318992654,0.8600200632764874,0.8729835294247004
4
+ 0,15000,0.867139109031403,0.1811354160308838,0.8167750946447605,0.10655976086854935,0.7798287478944413,0.8573964040435219,0.8630894015038928
5
+ 0,20000,0.8652457331421926,0.5867656469345093,0.810907497992261,0.042630940675735474,0.7694492552823,0.8570877382514083,0.8491018647322772
6
+ 0,25000,0.8699791728652186,0.19235554337501526,0.8189171463965605,0.11564554274082184,0.7757299647960241,0.8671965429431283,0.8779901051208325
7
+ 0,30000,0.8738470693246058,0.1700415164232254,0.8265335858953811,0.12816089391708374,0.7827916925412268,0.8754533528821669,0.880091749112012
8
+ 0,-1,0.8736036352817073,0.24405449628829956,0.8273541780947156,0.17230889201164246,0.7863200333657723,0.8729068600972297,0.8822842182205346
9
+ 1,5000,0.8749830948581321,0.18700861930847168,0.8258481907771007,0.1650511920452118,0.8056065164746459,0.847133266455745,0.8808449128337521
10
+ 1,10000,0.8773362906061508,0.056389499455690384,0.8286927441475971,0.05023118108510971,0.8076021678628973,0.8509144224091365,0.8806352233982097
11
+ 1,15000,0.8772551459251846,0.08140122890472412,0.8288128056914184,0.040914058685302734,0.7972057880105495,0.8630295547495949,0.8724934608981681
12
+ 1,20000,0.876146168618647,0.02961307018995285,0.8269898751016186,0.008039627224206924,0.7934482025101043,0.8634925534377652,0.8820982567462179
13
+ 1,25000,0.8766871331584215,0.018407132476568222,0.8281409975135293,0.006738942116498947,0.7869900618528042,0.8738328574735705,0.8403278253386028
14
+ 1,30000,0.8772551459251846,0.03199715167284012,0.8317026152111446,0.014339910820126534,0.7866382276042383,0.8822440003086658,0.8802174694024198
15
+ 1,-1,0.8785534608206432,0.03839807212352753,0.8313170836773521,0.012577023357152939,0.7772184185796751,0.8935103017208118,0.8790503615896275
README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Cross-Encoder for Quora Duplicate Questions Detection
2
+ This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
3
+
4
+ ## Training Data
5
+ This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
6
+
7
+ Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
8
+
9
+ ## Usage and Performance
10
+
11
+ Pre-trained models can be used like this:
12
+ ```
13
+ from sentence_transformers import CrossEncoder
14
+ model = CrossEncoder('model_name')
15
+ scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')])
16
+ ```
17
+
18
+ You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_num_labels": 3,
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "id2label": {
14
+ "0": "LABEL_0"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 4096,
18
+ "label2id": {
19
+ "LABEL_0": 0
20
+ },
21
+ "layer_norm_eps": 1e-05,
22
+ "max_position_embeddings": 514,
23
+ "model_type": "roberta",
24
+ "num_attention_heads": 16,
25
+ "num_hidden_layers": 24,
26
+ "pad_token_id": 1,
27
+ "type_vocab_size": 1,
28
+ "vocab_size": 50265
29
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a01c9ec97da7dcd86815855a2d91b43c96db6518c7da0f917bcbe00ced5d13e3
3
+ size 1421616585
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_max_length": 512, "special_tokens_map_file": "input-model/roberta-large-mnli/special_tokens_map.json", "full_tokenizer_file": null}
vocab.json ADDED
The diff for this file is too large to render. See raw diff