Create README
Browse files
README.md
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: no
|
3 |
+
license: CC-BY 4.0
|
4 |
+
thumbnail: https://raw.githubusercontent.com/NBAiLab/notram/master/images/nblogo_2.png
|
5 |
+
tags:
|
6 |
+
- nb-bert
|
7 |
+
- text-classification
|
8 |
+
- pytorch
|
9 |
+
- tensorflow
|
10 |
+
- norwegian
|
11 |
+
- bert
|
12 |
+
datasets:
|
13 |
+
- mnli
|
14 |
+
- multi_nli
|
15 |
+
- xnli
|
16 |
+
pipeline_tag: zero-shot-classification
|
17 |
+
widget:
|
18 |
+
- text: "Såframt uforutsette ting ikke inntreffer, kommer statsministeren til å presentere en plan for gjenåpningen av samfunnet før påske. Det bekrefter hun overfor Dagbladet."
|
19 |
+
candidate_labels: "politikk, helse, sport, religion, geografi"
|
20 |
+
multi_class: true
|
21 |
+
---
|
22 |
+
|
23 |
+
**Release 1.0** (March 11, 2021)
|
24 |
+
|
25 |
+
# NB-Bert base model finetuned on Norwegian machine translated MNLI
|
26 |
+
|
27 |
+
## Description
|
28 |
+
|
29 |
+
This finetuned model was created to show that it is possible to use a MNLI finetuned model for zero-shot classification. It is an alternative to finetuning a model on a specific annotated dataset.
|
30 |
+
|
31 |
+
## Intended use & limitations
|
32 |
+
|
33 |
+
## Training data
|
34 |
+
|
35 |
+
## More information
|
36 |
+
|
37 |
+
For more information on the model, see
|
38 |
+
|
39 |
+
https://github.com/NBAiLab/notram
|