Basti8499 commited on
Commit
5aec005
1 Parent(s): c514e65

Updates README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -15
README.md CHANGED
@@ -5,14 +5,13 @@ tags:
5
  - sentence-transformers
6
  - feature-extraction
7
  - sentence-similarity
8
-
 
9
  ---
10
 
11
  # {MODEL_NAME}
12
 
13
- This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 1024 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
-
15
- <!--- Describe your model here -->
16
 
17
  ## Usage (Sentence-Transformers)
18
 
@@ -33,15 +32,6 @@ embeddings = model.encode(sentences)
33
  print(embeddings)
34
  ```
35
 
36
-
37
-
38
- ## Evaluation Results
39
-
40
- <!--- Describe how your model was evaluated -->
41
-
42
- For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
43
-
44
-
45
  ## Training
46
  The model was trained with the parameters:
47
 
@@ -88,5 +78,4 @@ SentenceTransformer(
88
  ```
89
 
90
  ## Citing & Authors
91
-
92
- <!--- Describe where people can find more information -->
 
5
  - sentence-transformers
6
  - feature-extraction
7
  - sentence-similarity
8
+ - embeddings
9
+ license: mit
10
  ---
11
 
12
  # {MODEL_NAME}
13
 
14
+ This is a fine-tuned embedding model of [bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5). It was fine-tuned on a dataset based on an ISO 27001 text corpus consisting of text chunks (1024 characters) and associated questions. A total of 2.000 chunk and question pairs were generated. The fine-tuning process is specialized on an Information Retrieval task in which the generated questions are used to find the relevant chunks. The effectiveness of the model is evaluated on whether the correct chunk was retrieved, and the loss is calculated with the multiple negative ranking loss.
 
 
15
 
16
  ## Usage (Sentence-Transformers)
17
 
 
32
  print(embeddings)
33
  ```
34
 
 
 
 
 
 
 
 
 
 
35
  ## Training
36
  The model was trained with the parameters:
37
 
 
78
  ```
79
 
80
  ## Citing & Authors
81
+ Based on https://huggingface.co/BAAI/bge-large-en-v1.5 from Xiao et al. (2023) (C-Pack: Packaged Resources To Advance General Chinese Embedding)