Update README.md
Browse files
README.md
CHANGED
@@ -20,4 +20,38 @@ widget:
|
|
20 |
|
21 |
# Multilingual parliament sentiment regression model XLM-R-Parla-Sent
|
22 |
|
23 |
-
This model is
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
# Multilingual parliament sentiment regression model XLM-R-Parla-Sent
|
22 |
|
23 |
+
This model is based on [xlm-r-parla](https://huggingface.co/classla/xlm-r-parla) and fine-tuned on manually annotated sentiment datasets from United Kingdom, Czechia, Slovakia, Slovenia, Bosnia and Herzegovina, Croatia, and Serbia.
|
24 |
+
|
25 |
+
## Annotation schema
|
26 |
+
|
27 |
+
The discrete labels, present in the original dataset, were mapped to integers as follows:
|
28 |
+
|
29 |
+
```
|
30 |
+
"Negative": 0.0,
|
31 |
+
"M_Negative": 1.0,
|
32 |
+
"N_Neutral": 2.0,
|
33 |
+
"P_Neutral": 3.0,
|
34 |
+
"M_Positive": 4.0,
|
35 |
+
"Positive": 5.0,
|
36 |
+
```
|
37 |
+
Model was then fine-tuned on numeric labels and setup as regressor.
|
38 |
+
|
39 |
+
|
40 |
+
|
41 |
+
## Finetuning procedure
|
42 |
+
|
43 |
+
The fine-tuning procedure is described in this paper (ARXIV SUBMISSION to be added). Presumed optimal hyperparameters used are
|
44 |
+
```
|
45 |
+
num_train_epochs=4,
|
46 |
+
train_batch_size=32,
|
47 |
+
learning_rate=8e-6,
|
48 |
+
regression=True
|
49 |
+
```
|
50 |
+
|
51 |
+
## Results
|
52 |
+
|
53 |
+
Results reported were obtained from 10 fine-tuning runs.
|
54 |
+
|
55 |
+
test dataset | R^2
|
56 |
+
BCS | 0.6146 ± 0.0104
|
57 |
+
EN | 0.6722 ± 0.0100
|