roberta-argument / README.md
chkla's picture
Update README.md
90d2663
|
raw
history blame
No virus
1.82 kB
Welcome to **RoBERTArg**!
πŸ€– **Model description**:
This model was trained on ~40k heterogeneous manually annotated sentences (πŸ“š Stab et al. 2018) of controversial topics (abortion etc.) to classify text into one of two labels: 🏷 **NON-ARGUMENT** (0) and **ARGUMENT** (1).
**Dataset**
Please note that the label distribution in the dataset is imbalanced:
* NON-ARGUMENTS:
* ARGUMENTS:
**Model training**
**RoBERTArg** was fine-tuned on a $RoBERTA_{base}$ pre-trained model using the HuggingFace trainer with the following hyperparameters. The hyperparameters were determined using a hyperparameter search on a 20% validation set.
```
training_args = TrainingArguments(
num_train_epochs=2,
learning_rate=2.3102e-06,
seed=8,
per_device_train_batch_size=64,
per_device_eval_batch_size=64,
)
```
**Evaluation**
The model was evaluated using 20% of the sentences (80-20 train-test split).
| Model | Acc | F1 | R arg | R non | P arg | P non |
|----|----|----|----|----|----|----|
| RoBERTArg | 0.8193 | 0.8021 | 0.8463 | 0.7986 | 0.7623 | 0.8719 |
Showing the **confusion matrix** using the 20% of the sentences as an evaluation set:
| | ARGUMENT | NON-ARGUMENT |
|----|----|----|
| ARGUMENT | 2213 | 558 |
| NON-ARGUMENT | 325 | 1790 |
**Intended Uses & Potential Limitations**
The model can be a practical starting point to the complex topic **Argument Mining**. It is a quite challenging task due to the different conceptions of an argument.
This model is a part of an open-source project providing several models to detect arguments in text.
πŸ‘‰πŸΎ Check out _chkla/argument-analyzer/_ for more details.
Enjoy and stay tuned! πŸš€
πŸ“šStab et al. (2018) https://public.ukp.informatik.tu-darmstadt.de/UKP_Webpage/publications/2018/2018_EMNLP_CS_Cross-topicArgumentMining.pdf