Update README.md
Browse files
README.md
CHANGED
@@ -764,7 +764,7 @@ model-index:
|
|
764 |
---
|
765 |
|
766 |
# all-mpnet-base-v2-negation
|
767 |
-
This is a fine-tuned [sentence-transformers](https://www.SBERT.net) model to perform better
|
768 |
|
769 |
It maps sentences and paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
770 |
|
@@ -851,8 +851,5 @@ We used [`sentence-transformers/all-mpnet-base-v2`](https://huggingface.co/sente
|
|
851 |
We fine-tuned the model on the [CANNOT dataset](https://huggingface.co/datasets/tum-nlp/cannot-dataset) using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch. We then apply the cross entropy loss by comparing with true pairs.
|
852 |
|
853 |
#### Hyper parameters
|
854 |
-
We followed an analogous approach to [how other Sentence Transformers were trained](https://github.com/UKPLab/sentence-transformers/blob/3e1929fddef16df94f8bc6e3b10598a98f46e62d/examples/training/nli/training_nli_v2.py).
|
855 |
-
|
856 |
-
We took the first 90% of samples from the CANNOT dataset as the training split.
|
857 |
-
|
858 |
We used a batch size of 64 and trained for 1 epoch.
|
|
|
764 |
---
|
765 |
|
766 |
# all-mpnet-base-v2-negation
|
767 |
+
**This is a fine-tuned [sentence-transformers](https://www.SBERT.net) model to perform better on negated pairs of sentences.**
|
768 |
|
769 |
It maps sentences and paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
770 |
|
|
|
851 |
We fine-tuned the model on the [CANNOT dataset](https://huggingface.co/datasets/tum-nlp/cannot-dataset) using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch. We then apply the cross entropy loss by comparing with true pairs.
|
852 |
|
853 |
#### Hyper parameters
|
854 |
+
We followed an analogous approach to [how other Sentence Transformers were trained](https://github.com/UKPLab/sentence-transformers/blob/3e1929fddef16df94f8bc6e3b10598a98f46e62d/examples/training/nli/training_nli_v2.py). We took the first 90% of samples from the CANNOT dataset as the training split.
|
|
|
|
|
|
|
855 |
We used a batch size of 64 and trained for 1 epoch.
|