5roop commited on
Commit
6ca647b
1 Parent(s): 843b7e8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -20,4 +20,38 @@ widget:
20
 
21
  # Multilingual parliament sentiment regression model XLM-R-Parla-Sent
22
 
23
- This model is a result of a pre-training campaign, which trained XLM-Roberta-large on parliamentary transcriptions (collected in ParlaMint and EuroParl corpora), and fine-tuned on manually annotated sentiment datasets from United Kingdom, Czechia, Slovakia, Slovenia, Bosnia and Herzegovina, Croatia, and Serbia.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  # Multilingual parliament sentiment regression model XLM-R-Parla-Sent
22
 
23
+ This model is based on [xlm-r-parla](https://huggingface.co/classla/xlm-r-parla) and fine-tuned on manually annotated sentiment datasets from United Kingdom, Czechia, Slovakia, Slovenia, Bosnia and Herzegovina, Croatia, and Serbia.
24
+
25
+ ## Annotation schema
26
+
27
+ The discrete labels, present in the original dataset, were mapped to integers as follows:
28
+
29
+ ```
30
+ "Negative": 0.0,
31
+ "M_Negative": 1.0,
32
+ "N_Neutral": 2.0,
33
+ "P_Neutral": 3.0,
34
+ "M_Positive": 4.0,
35
+ "Positive": 5.0,
36
+ ```
37
+ Model was then fine-tuned on numeric labels and setup as regressor.
38
+
39
+
40
+
41
+ ## Finetuning procedure
42
+
43
+ The fine-tuning procedure is described in this paper (ARXIV SUBMISSION to be added). Presumed optimal hyperparameters used are
44
+ ```
45
+ num_train_epochs=4,
46
+ train_batch_size=32,
47
+ learning_rate=8e-6,
48
+ regression=True
49
+ ```
50
+
51
+ ## Results
52
+
53
+ Results reported were obtained from 10 fine-tuning runs.
54
+
55
+ test dataset | R^2
56
+ BCS | 0.6146 ± 0.0104
57
+ EN | 0.6722 ± 0.0100