|
--- |
|
license: cc-by-nc-4.0 |
|
--- |
|
|
|
## Model Specification |
|
- This is the **Republican** community GPT-2 language model, fine-tuned on 4.7M (~100M tokens) tweets of Republican Twitter users between 2019-01-01 and 2020-04-10. |
|
- For more details about the `CommunityLM` project, please refer to this [our paper](https://arxiv.org/abs/2209.07065) and [github](https://github.com/hjian42/communitylm) page. |
|
- In the paper, it is referred as the `Fine-tuned CommunityLM` for the Republican Twitter community. |
|
|
|
## How to use the model |
|
|
|
- **PRE-PROCESSING**: when you apply the model on tweets, please make sure that tweets are preprocessed by the [TweetTokenizer](https://github.com/VinAIResearch/BERTweet/blob/master/TweetNormalizer.py) to get the best performance. |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("CommunityLM/republican-twitter-gpt2") |
|
|
|
model = AutoModelForCausalLM.from_pretrained("CommunityLM/republican-twitter-gpt2") |
|
``` |
|
|
|
## References |
|
|
|
If you use this repository in your research, please kindly cite [our paper](https://arxiv.org/abs/2209.07065): |
|
|
|
```bibtex |
|
@inproceedings{jiang-etal-2022-communitylm, |
|
title = "CommunityLM: Probing Partisan Worldviews from Language Models", |
|
author = {Jiang, Hang and Beeferman, Doug and Roy, Brandon and Roy, Deb}, |
|
booktitle = "Proceedings of the 29th International Conference on Computational Linguistics", |
|
year = "2022", |
|
publisher = "International Committee on Computational Linguistics", |
|
} |
|
``` |