hiiamsid commited on
Commit
31c4ddf
1 Parent(s): e1320ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -7
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - transformers
8
  ---
9
 
10
- # {MODEL_NAME}
11
 
12
  This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
13
 
@@ -27,7 +27,7 @@ Then you can use the model like this:
27
  from sentence_transformers import SentenceTransformer
28
  sentences = ["This is an example sentence", "Each sentence is converted"]
29
 
30
- model = SentenceTransformer('{MODEL_NAME}')
31
  embeddings = model.encode(sentences)
32
  print(embeddings)
33
  ```
@@ -53,8 +53,8 @@ def mean_pooling(model_output, attention_mask):
53
  sentences = ['This is an example sentence', 'Each sentence is converted']
54
 
55
  # Load model from HuggingFace Hub
56
- tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
57
- model = AutoModel.from_pretrained('{MODEL_NAME}')
58
 
59
  # Tokenize sentences
60
  encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
@@ -73,10 +73,15 @@ print(sentence_embeddings)
73
 
74
 
75
  ## Evaluation Results
 
 
 
 
 
 
 
76
 
77
- <!--- Describe how your model was evaluated -->
78
-
79
- For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
80
 
81
 
82
  ## Training
 
7
  - transformers
8
  ---
9
 
10
+ # hiiamsid/sentence_similarity_spanish_es
11
 
12
  This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
13
 
 
27
  from sentence_transformers import SentenceTransformer
28
  sentences = ["This is an example sentence", "Each sentence is converted"]
29
 
30
+ model = SentenceTransformer('hiiamsid/sentence_similarity_spanish_es')
31
  embeddings = model.encode(sentences)
32
  print(embeddings)
33
  ```
 
53
  sentences = ['This is an example sentence', 'Each sentence is converted']
54
 
55
  # Load model from HuggingFace Hub
56
+ tokenizer = AutoTokenizer.from_pretrained('hiiamsid/sentence_similarity_spanish_es')
57
+ model = AutoModel.from_pretrained('hiiamsid/sentence_similarity_spanish_es')
58
 
59
  # Tokenize sentences
60
  encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
 
73
 
74
 
75
  ## Evaluation Results
76
+ ```
77
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
78
+ 0,-1,0.8441511943881872,0.842139456614267,0.8079280123390863,0.8133773844408285,0.8061045432939875,0.8120313077224331,0.765564331583017,0.7758911212599943
79
+ 1,-1,0.8505781800765124,0.8483036121137681,0.824043571251859,0.828168962987315,0.8227441143847719,0.8269561222926696,0.7895885844553006,0.7956808786541772
80
+ 2,-1,0.8540176327819012,0.8524414878771152,0.8278950073844416,0.8323727470271813,0.8268352374519953,0.8316947483386466,0.8027944039984911,0.8062262061975706
81
+ 3,-1,0.8532432895679434,0.851780980304441,0.8298723338734442,0.8340725567708174,0.8289668426348045,0.8333822263715207,0.8043286320186501,0.8063301333153703
82
+ ```
83
 
84
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=hiiamsid/sentence_similarity_spanish_es)
 
 
85
 
86
 
87
  ## Training