Climate-TwitterBERT's picture
Update README.md
9fa04dc
|
raw
history blame
2.14 kB
# Model Card Climate-TwitterBERT-step-1
## Overview:
Using Covid-Twitter-BERT-v2 (https://huggingface.co/digitalepidemiologylab/covid-twitter-bert-v2) as the starting model, we continued domain-adaptive pre-training on a corpus of firm tweets between 2007 and 2020. The model was then fine-tuned on the downstream task to classify whether a given tweet is related to climate change topics.
The model provides a label and probability score, indicating whether a given tweet is related to climate change topics (label = 1) or not (label = 0).
## Performance metrics:
Based on the test set, the model achieves the following results:
• Loss: 0.0632
• F1-weighted: 0.9778
• F1: 0.9148
• Accuracy: 0.9775
• Precision: 0. 8841
• Recall: 0. 9477
## Example usage:
```python
from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification
task_name = 'binary'
model_name = Climate-TwitterBERT/ Climate-TwitterBERT-step1'
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)
pipe = pipeline(task=‘binary‘, model=model, tokenizer=tokenizer)
tweet = "We are committed to significantly cutting our carbon emissions by 30% before 2030."
result = pipe(tweet)
# The 'result' variable will contain the classification output: 0 = non-climate tweet, 1= climate tweet
```
## Citation:
```bibtex
@article{fzz2023climatetwitter,
title={Responding to Climate Change crisis - firms' tradeoffs},
author={Fritsch, Felix and Zhang, Qi and Zheng, Xiang},
journal={Working paper},
year={2023},
institution={University of Mannheim, the Chinese University of Hong Kong, and NHH Norwegian School of Economics},
url={https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4527255}
}
```
Fritsch, F., Zhang, Q., & Zheng, X. (2023). Responding to Climate Change crisis - firms' tradeoffs [Working paper]. University of Mannheim, the Chinese University of Hong Kong, and NHH Norwegian School of Economics.
## Framework versions
• Transformers 4.28.1
• Pytorch 2.0.1+cu118
• Datasets 2.14.1
• Tokenizers 0.13.3