dmlls commited on
Commit
1d1c608
1 Parent(s): 4dfaede

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -764,7 +764,7 @@ model-index:
764
  ---
765
 
766
  # all-mpnet-base-v2-negation
767
- This is a fine-tuned [sentence-transformers](https://www.SBERT.net) model to perform better with negated pairs of sentences.
768
 
769
  It maps sentences and paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
770
 
@@ -851,8 +851,5 @@ We used [`sentence-transformers/all-mpnet-base-v2`](https://huggingface.co/sente
851
  We fine-tuned the model on the [CANNOT dataset](https://huggingface.co/datasets/tum-nlp/cannot-dataset) using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch. We then apply the cross entropy loss by comparing with true pairs.
852
 
853
  #### Hyper parameters
854
- We followed an analogous approach to [how other Sentence Transformers were trained](https://github.com/UKPLab/sentence-transformers/blob/3e1929fddef16df94f8bc6e3b10598a98f46e62d/examples/training/nli/training_nli_v2.py).
855
-
856
- We took the first 90% of samples from the CANNOT dataset as the training split.
857
-
858
  We used a batch size of 64 and trained for 1 epoch.
 
764
  ---
765
 
766
  # all-mpnet-base-v2-negation
767
+ **This is a fine-tuned [sentence-transformers](https://www.SBERT.net) model to perform better on negated pairs of sentences.**
768
 
769
  It maps sentences and paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
770
 
 
851
  We fine-tuned the model on the [CANNOT dataset](https://huggingface.co/datasets/tum-nlp/cannot-dataset) using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch. We then apply the cross entropy loss by comparing with true pairs.
852
 
853
  #### Hyper parameters
854
+ We followed an analogous approach to [how other Sentence Transformers were trained](https://github.com/UKPLab/sentence-transformers/blob/3e1929fddef16df94f8bc6e3b10598a98f46e62d/examples/training/nli/training_nli_v2.py). We took the first 90% of samples from the CANNOT dataset as the training split.
 
 
 
855
  We used a batch size of 64 and trained for 1 epoch.