Update README.md
Browse files
README.md
CHANGED
@@ -39,7 +39,7 @@ When the TSystems model was published, only the STSb dataset was used for STS tr
|
|
39 |
Therefore it is included in our model, but expanded to include SICK and Priya22 semantic textual relatedness:
|
40 |
- SICK was partly used in STSb, but our custom translation using [DeepL](https://www.deepl.com/) leads to slightly different phrases. This approach allows more examples to be included in the training.
|
41 |
- The Priya22 semantic textual relatedness dataset published in 2022 was also translated into German via DeepL and added to the training data. Since it does not have a train-test-split, it was created independently at a ratio of 80:20.
|
42 |
-
|
43 |
All training and test data (STSb, Sick, Priya22) were checked for duplicates within and with each other and removed if found.
|
44 |
Because the test data is prioritized, duplicated entries between test-train are exclusively removed from train split.
|
45 |
The final used datasets can be viewed here: [datasets_sts_paraphrase_xlm-roberta-base_de-en](https://gitlab.com/sense.ai.tion-public/datasets_sts_paraphrase_xlm-roberta-base_de-en)
|
@@ -74,7 +74,7 @@ In addition, the test samples used are evaluated individually for each data set
|
|
74 |
This subdivision per data set allows for a fair overall assessment, since external models are not built on the same data basis as the model presented here.
|
75 |
The data is not evenly distributed in either training or testing!
|
76 |
|
77 |
-
**❗Some models are only usable for one language (because they are monolingual). They will almost not perform at all in the other two tables
|
78 |
|
79 |
The first table shows the evaluation results for **cross-lingual (German-English-Mixed)** based on _Spearman_:
|
80 |
**model**|**STSb**|**SICK**|**Priya22**|**all**|
|
|
|
39 |
Therefore it is included in our model, but expanded to include SICK and Priya22 semantic textual relatedness:
|
40 |
- SICK was partly used in STSb, but our custom translation using [DeepL](https://www.deepl.com/) leads to slightly different phrases. This approach allows more examples to be included in the training.
|
41 |
- The Priya22 semantic textual relatedness dataset published in 2022 was also translated into German via DeepL and added to the training data. Since it does not have a train-test-split, it was created independently at a ratio of 80:20.
|
42 |
+
|
43 |
All training and test data (STSb, Sick, Priya22) were checked for duplicates within and with each other and removed if found.
|
44 |
Because the test data is prioritized, duplicated entries between test-train are exclusively removed from train split.
|
45 |
The final used datasets can be viewed here: [datasets_sts_paraphrase_xlm-roberta-base_de-en](https://gitlab.com/sense.ai.tion-public/datasets_sts_paraphrase_xlm-roberta-base_de-en)
|
|
|
74 |
This subdivision per data set allows for a fair overall assessment, since external models are not built on the same data basis as the model presented here.
|
75 |
The data is not evenly distributed in either training or testing!
|
76 |
|
77 |
+
**❗Some models are only usable for one language (because they are monolingual). They will almost not perform at all in the other two tables. Still, they are good models in certain applications❗**
|
78 |
|
79 |
The first table shows the evaluation results for **cross-lingual (German-English-Mixed)** based on _Spearman_:
|
80 |
**model**|**STSb**|**SICK**|**Priya22**|**all**|
|