sberbank-ai
commited on
Commit
•
3e18da9
1
Parent(s):
ad8f57a
Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ def mean_pooling(model_output, attention_mask):
|
|
21 |
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
|
22 |
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
23 |
sum_embeddings = torch.sum(token_embeddings * input_mask_expanded, 1)
|
24 |
-
sum_mask = torch.clamp(input_mask_expanded.sum(1), min=1e-
|
25 |
return sum_embeddings / sum_mask
|
26 |
#Sentences we want sentence embeddings for
|
27 |
sentences = ['Привет! Как твои дела?',
|
|
|
21 |
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
|
22 |
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
23 |
sum_embeddings = torch.sum(token_embeddings * input_mask_expanded, 1)
|
24 |
+
sum_mask = torch.clamp(input_mask_expanded.sum(1), min=1e-10)
|
25 |
return sum_embeddings / sum_mask
|
26 |
#Sentences we want sentence embeddings for
|
27 |
sentences = ['Привет! Как твои дела?',
|