Update README.md
Browse files
README.md
CHANGED
@@ -28,15 +28,17 @@ The learning objective for FSP is to predict the index of the correct label.
|
|
28 |
A cross-entropy loss is used for tuning the model.
|
29 |
|
30 |
## Model variations
|
31 |
-
There are
|
32 |
|
33 |
-
| Model | Backbone | #params |
|
34 |
-
|
35 |
-
| [zero-shot-classify-SSTuning-base](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-base) | [roberta-base](https://huggingface.co/roberta-base) | 125M |
|
36 |
-
| [zero-shot-classify-SSTuning-large](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-large) | [roberta-large](https://huggingface.co/roberta-large) | 355M | Medium | Medium | 5.12M |
|
37 |
-
| [zero-shot-classify-SSTuning-ALBERT](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-ALBERT) | [albert-xxlarge-v2](https://huggingface.co/albert-xxlarge-v2) | 235M |
|
|
|
38 |
|
39 |
-
Please note that zero-shot-classify-SSTuning-
|
|
|
40 |
|
41 |
|
42 |
## Intended uses & limitations
|
|
|
28 |
A cross-entropy loss is used for tuning the model.
|
29 |
|
30 |
## Model variations
|
31 |
+
There are four versions of models released. The details are:
|
32 |
|
33 |
+
| Model | Backbone | #params | lang | acc | Speed | #Train
|
34 |
+
|------------|-----------|----------|-------|-------|----|-------------|
|
35 |
+
| [zero-shot-classify-SSTuning-base](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-base) | [roberta-base](https://huggingface.co/roberta-base) | 125M | En | Low | High | 20.48M |
|
36 |
+
| [zero-shot-classify-SSTuning-large](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-large) | [roberta-large](https://huggingface.co/roberta-large) | 355M | En | Medium | Medium | 5.12M |
|
37 |
+
| [zero-shot-classify-SSTuning-ALBERT](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-ALBERT) | [albert-xxlarge-v2](https://huggingface.co/albert-xxlarge-v2) | 235M | En | High | Low| 5.12M |
|
38 |
+
| [zero-shot-classify-SSTuning-XLM-R](https://huggingface.co/DAMO-NLP-SG/zero-shot-classify-SSTuning-XLM-R) | [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) | 278M | Multi | - | - | 20.48M |
|
39 |
|
40 |
+
Please note that zero-shot-classify-SSTuning-XLM-R is trained with 20.48M English samples only. However, it can also be used in other languages as long as xlm-roberta supports.
|
41 |
+
Please check [this repository](https://github.com/DAMO-NLP-SG/SSTuning) for the performance of each model.
|
42 |
|
43 |
|
44 |
## Intended uses & limitations
|