Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ license: apache-2.0
|
|
10 |
# Please use 'Roberta' related functions to load this model!
|
11 |
|
12 |
This repository contains the resources in our paper
|
13 |
-
**[
|
14 |
*Ramon Villa-Cox, Evan Williams, Kathleen M. Carley*
|
15 |
|
16 |
We pre-trained a BERT language model, we call *TwBETO_v0* following the robust approach introduced in RoBERTa. We opted for the smaller architecture dimensions introduced in DistilBERT, namely, 6 hidden layers with 12 attention heads. We also reduce the model's maximum sequence length to 128 tokens, following another BERT instantiation trained on English Twitter data (*BERTweet*). We utilize the RoBERTa implementation in the Hugging Face library and optimize the model using Adam with weight decay, a linear schedule with warmup and a maximum learning rate of 2e-4. We use a global batch size (via gradient accumulation) of 5k across 4 Titan XP GPUs (12 GB RAM each) and trained the model for 650 hours.
|
|
|
10 |
# Please use 'Roberta' related functions to load this model!
|
11 |
|
12 |
This repository contains the resources in our paper
|
13 |
+
**[Social Context in Political Stance Detection: Impact and Extrapolation]**
|
14 |
*Ramon Villa-Cox, Evan Williams, Kathleen M. Carley*
|
15 |
|
16 |
We pre-trained a BERT language model, we call *TwBETO_v0* following the robust approach introduced in RoBERTa. We opted for the smaller architecture dimensions introduced in DistilBERT, namely, 6 hidden layers with 12 attention heads. We also reduce the model's maximum sequence length to 128 tokens, following another BERT instantiation trained on English Twitter data (*BERTweet*). We utilize the RoBERTa implementation in the Hugging Face library and optimize the model using Adam with weight decay, a linear schedule with warmup and a maximum learning rate of 2e-4. We use a global batch size (via gradient accumulation) of 5k across 4 Titan XP GPUs (12 GB RAM each) and trained the model for 650 hours.
|