Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ tags:
|
|
23 |
## Model Description
|
24 |
|
25 |
This model consists of a fine-tuned version of google-bert/bert-base-cased for a propaganda detection task. It is effectively a binary classifier, determining wether propaganda is present in the output string.
|
26 |
-
This model was created by [`Identrics`](https://identrics.ai/), in the scope of the
|
27 |
|
28 |
|
29 |
## Uses
|
@@ -55,14 +55,22 @@ print(output.logits)
|
|
55 |
The training datasets for the model consist of a balanced set totaling 840 English examples that include both propaganda and non-propaganda content. These examples are collected from a variety of traditional media and social media sources, ensuring a diverse range of content. Aditionally, the training dataset is enriched with AI-generated samples. The total distribution of the training data is shown in the table below:
|
56 |
|
57 |
|
58 |
-
![image/png](https://cdn-uploads.huggingface.co/production/uploads/66741cdd8123010b8f63f965/KyUIrMGWmmpnE67WZeQaN.png)
|
59 |
|
|
|
60 |
|
61 |
-
The model was then tested on a smaller evaluation dataset, achieving an f1 score of 0.807. The evaluation dataset is distributed as such:
|
62 |
|
|
|
63 |
|
|
|
64 |
|
65 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
|
67 |
|
68 |
|
|
|
23 |
## Model Description
|
24 |
|
25 |
This model consists of a fine-tuned version of google-bert/bert-base-cased for a propaganda detection task. It is effectively a binary classifier, determining wether propaganda is present in the output string.
|
26 |
+
This model was created by [`Identrics`](https://identrics.ai/), in the scope of the WASPer project.
|
27 |
|
28 |
|
29 |
## Uses
|
|
|
55 |
The training datasets for the model consist of a balanced set totaling 840 English examples that include both propaganda and non-propaganda content. These examples are collected from a variety of traditional media and social media sources, ensuring a diverse range of content. Aditionally, the training dataset is enriched with AI-generated samples. The total distribution of the training data is shown in the table below:
|
56 |
|
57 |
|
|
|
58 |
|
59 |
+
The model was then tested on a smaller evaluation dataset, achieving an f1 score of 0.807.
|
60 |
|
|
|
61 |
|
62 |
+
## Citation
|
63 |
|
64 |
+
If you find our work useful, please consider citing WASPer:
|
65 |
|
66 |
+
```
|
67 |
+
@article{bai2024longwriter,
|
68 |
+
title={LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs},
|
69 |
+
author={Yushi Bai and Jiajie Zhang and Xin Lv and Linzhi Zheng and Siqi Zhu and Lei Hou and Yuxiao Dong and Jie Tang and Juanzi Li},
|
70 |
+
journal={arXiv preprint arXiv:2408.07055},
|
71 |
+
year={2024}
|
72 |
+
}
|
73 |
+
```
|
74 |
|
75 |
|
76 |
|