Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
---
|
2 |
-
pipeline_tag: sentence-similarity
|
3 |
tags:
|
4 |
- sentence-transformers
|
5 |
- feature-extraction
|
@@ -10,6 +9,9 @@ datasets:
|
|
10 |
license: cc-by-sa-4.0
|
11 |
language:
|
12 |
- ja
|
|
|
|
|
|
|
13 |
---
|
14 |
|
15 |
# unsup-simcse-ja-large
|
@@ -77,6 +79,24 @@ SentenceTransformer(
|
|
77 |
)
|
78 |
```
|
79 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
## Citing & Authors
|
81 |
|
82 |
```
|
|
|
1 |
---
|
|
|
2 |
tags:
|
3 |
- sentence-transformers
|
4 |
- feature-extraction
|
|
|
9 |
license: cc-by-sa-4.0
|
10 |
language:
|
11 |
- ja
|
12 |
+
metrics:
|
13 |
+
- spearmanr
|
14 |
+
library_name: sentence-transformers
|
15 |
---
|
16 |
|
17 |
# unsup-simcse-ja-large
|
|
|
79 |
)
|
80 |
```
|
81 |
|
82 |
+
## Model Summary
|
83 |
+
|
84 |
+
- Fine-tuning method: Unsupervised SimCSE
|
85 |
+
- Base model: [cl-tohoku/bert-large-japanese-v2](https://huggingface.co/cl-tohoku/bert-large-japanese-v2)
|
86 |
+
- Training dataset: [Wiki40B](https://huggingface.co/datasets/wiki40b)
|
87 |
+
- Pooling strategy: cls (with an extra MLP layer only during training)
|
88 |
+
- Hidden size: 1024
|
89 |
+
- Learning rate: 3e-5
|
90 |
+
- Batch size: 64
|
91 |
+
- Temperature: 0.05
|
92 |
+
- Max sequence length: 64
|
93 |
+
- Number of training examples: 2^20
|
94 |
+
- Validation interval (steps): 2^6
|
95 |
+
- Warmup ratio: 0.1
|
96 |
+
- Dtype: BFloat16
|
97 |
+
|
98 |
+
See the [GitHub repository](https://github.com/hppRC/simple-simcse-ja) for a detailed experimental setup.
|
99 |
+
|
100 |
## Citing & Authors
|
101 |
|
102 |
```
|