Nikola299 commited on
Commit
7a81dc3
1 Parent(s): a3b8f57

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -4
README.md CHANGED
@@ -23,7 +23,7 @@ tags:
23
  ## Model Description
24
 
25
  This model consists of a fine-tuned version of google-bert/bert-base-cased for a propaganda detection task. It is effectively a binary classifier, determining wether propaganda is present in the output string.
26
- This model was created by [`Identrics`](https://identrics.ai/), in the scope of the Wasper project.
27
 
28
 
29
  ## Uses
@@ -55,14 +55,22 @@ print(output.logits)
55
  The training datasets for the model consist of a balanced set totaling 840 English examples that include both propaganda and non-propaganda content. These examples are collected from a variety of traditional media and social media sources, ensuring a diverse range of content. Aditionally, the training dataset is enriched with AI-generated samples. The total distribution of the training data is shown in the table below:
56
 
57
 
58
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66741cdd8123010b8f63f965/KyUIrMGWmmpnE67WZeQaN.png)
59
 
 
60
 
61
- The model was then tested on a smaller evaluation dataset, achieving an f1 score of 0.807. The evaluation dataset is distributed as such:
62
 
 
63
 
 
64
 
65
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66741cdd8123010b8f63f965/5MOK5L7Tq9Ff64t0rPo17.png)
 
 
 
 
 
 
 
66
 
67
 
68
 
 
23
  ## Model Description
24
 
25
  This model consists of a fine-tuned version of google-bert/bert-base-cased for a propaganda detection task. It is effectively a binary classifier, determining wether propaganda is present in the output string.
26
+ This model was created by [`Identrics`](https://identrics.ai/), in the scope of the WASPer project.
27
 
28
 
29
  ## Uses
 
55
  The training datasets for the model consist of a balanced set totaling 840 English examples that include both propaganda and non-propaganda content. These examples are collected from a variety of traditional media and social media sources, ensuring a diverse range of content. Aditionally, the training dataset is enriched with AI-generated samples. The total distribution of the training data is shown in the table below:
56
 
57
 
 
58
 
59
+ The model was then tested on a smaller evaluation dataset, achieving an f1 score of 0.807.
60
 
 
61
 
62
+ ## Citation
63
 
64
+ If you find our work useful, please consider citing WASPer:
65
 
66
+ ```
67
+ @article{bai2024longwriter,
68
+ title={LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs},
69
+ author={Yushi Bai and Jiajie Zhang and Xin Lv and Linzhi Zheng and Siqi Zhu and Lei Hou and Yuxiao Dong and Jie Tang and Juanzi Li},
70
+ journal={arXiv preprint arXiv:2408.07055},
71
+ year={2024}
72
+ }
73
+ ```
74
 
75
 
76