|
|
|
--- |
|
--- |
|
license: mit |
|
language: |
|
- fr |
|
tags: |
|
- zero-shot-classification |
|
- text-classification |
|
- pytorch |
|
metrics: |
|
- accuracy |
|
- f1-score |
|
--- |
|
# xlm-roberta-large-french-execorder-cap-v3 |
|
## Model description |
|
An `xlm-roberta-large` model finetuned on french training data containing texts of the `execorder` domain labelled with [major topic codes](https://www.comparativeagendas.net/pages/master-codebook) from the [Comparative Agendas Project](https://www.comparativeagendas.net/). |
|
|
|
## How to use the model |
|
#### Loading and tokenizing input data |
|
```python |
|
import pandas as pd |
|
import numpy as np |
|
from datasets import Dataset |
|
from transformers import (AutoModelForSequenceClassification, AutoTokenizer, |
|
Trainer, TrainingArguments) |
|
|
|
CAP_NUM_DICT = {0: '1', 1: '2', 2: '3', 3: '4', 4: '5', 5: '6', |
|
6: '7', 7: '8', 8: '9', 9: '10', 10: '12', 11: '13', 12: '14', |
|
13: '15', 14: '16', 15: '17', 16: '18', 17: '19', 18: '20', 19: |
|
'21', 20: '23', 21: '999'} |
|
|
|
tokenizer = AutoTokenizer.from_pretrained('xlm-roberta-large') |
|
num_labels = len(CAP_NUM_DICT) |
|
|
|
def tokenize_dataset(data : pd.DataFrame): |
|
tokenized = tokenizer(data["text"], |
|
max_length=MAXLEN, |
|
truncation=True, |
|
padding="max_length") |
|
return tokenized |
|
|
|
hg_data = Dataset.from_pandas(data) |
|
dataset = hg_data.map(tokenize_dataset, batched=True, remove_columns=hg_data.column_names) |
|
``` |
|
|
|
#### Inference using the Trainer class |
|
```python |
|
model = AutoModelForSequenceClassification.from_pretrained('poltextlab/xlm-roberta-large-french-execorder-cap-v3', |
|
num_labels=21, |
|
problem_type="multi_label_classification") ) |
|
|
|
training_args = TrainingArguments( |
|
output_dir='.', |
|
per_device_train_batch_size=8, |
|
per_device_eval_batch_size=8 |
|
) |
|
|
|
trainer = Trainer( |
|
model=model, |
|
args=training_args |
|
) |
|
|
|
probs = trainer.predict(test_dataset=dataset).predictions |
|
predicted = pd.DataFrame(np.argmax(probs, axis=1)).replace({0: CAP_NUM_DICT}).rename( |
|
columns={0: 'predicted'}).reset_index(drop=True) |
|
|
|
``` |
|
|
|
### Fine-tuning procedure |
|
`xlm-roberta-large-french-execorder-cap-v3` was fine-tuned using the Hugging Face Trainer class with the following hyperparameters: |
|
```python |
|
training_args = TrainingArguments( |
|
output_dir=f"../model/{model_dir}/tmp/", |
|
logging_dir=f"../logs/{model_dir}/", |
|
logging_strategy='epoch', |
|
num_train_epochs=10, |
|
per_device_train_batch_size=8, |
|
per_device_eval_batch_size=8, |
|
learning_rate=5e-06, |
|
seed=42, |
|
save_strategy='epoch', |
|
evaluation_strategy='epoch', |
|
save_total_limit=1, |
|
load_best_model_at_end=True |
|
) |
|
``` |
|
We also incorporated an EarlyStoppingCallback in the process with a patience of 2 epochs. |
|
|
|
## Model performance |
|
The model was evaluated on a test set of 494 examples (10% of the available data).<br> |
|
Model accuracy is **0.74**. |
|
| label | precision | recall | f1-score | support | |
|
|:-------------|------------:|---------:|-----------:|----------:| |
|
| 0 | 0.85 | 0.85 | 0.85 | 46 | |
|
| 1 | 0.89 | 0.47 | 0.62 | 17 | |
|
| 2 | 0.56 | 0.75 | 0.64 | 12 | |
|
| 3 | 0.79 | 0.65 | 0.71 | 17 | |
|
| 4 | 0.78 | 0.87 | 0.82 | 53 | |
|
| 5 | 0.76 | 0.84 | 0.8 | 19 | |
|
| 6 | 0.6 | 0.75 | 0.67 | 8 | |
|
| 7 | 0.64 | 0.88 | 0.74 | 8 | |
|
| 8 | 0.86 | 0.75 | 0.8 | 8 | |
|
| 9 | 0.7 | 0.86 | 0.78 | 22 | |
|
| 10 | 0.74 | 0.73 | 0.74 | 44 | |
|
| 11 | 0.33 | 0.33 | 0.33 | 3 | |
|
| 12 | 0.73 | 0.53 | 0.62 | 15 | |
|
| 13 | 0.7 | 0.77 | 0.73 | 39 | |
|
| 14 | 0.83 | 0.83 | 0.83 | 18 | |
|
| 15 | 1 | 0.14 | 0.25 | 7 | |
|
| 16 | 0 | 0 | 0 | 8 | |
|
| 17 | 0.66 | 0.83 | 0.73 | 48 | |
|
| 18 | 0.76 | 0.73 | 0.75 | 82 | |
|
| 19 | 0 | 0 | 0 | 6 | |
|
| 20 | 0.8 | 0.86 | 0.83 | 14 | |
|
| macro avg | 0.67 | 0.64 | 0.63 | 494 | |
|
| weighted avg | 0.73 | 0.74 | 0.73 | 494 | |
|
|
|
## Inference platform |
|
This model is used by the [CAP Babel Machine](https://babel.poltextlab.com), an open-source and free natural language processing tool, designed to simplify and speed up projects for comparative research. |
|
|
|
## Cooperation |
|
Model performance can be significantly improved by extending our training sets. We appreciate every submission of CAP-coded corpora (of any domain and language) at poltextlab{at}poltextlab{dot}com or by using the [CAP Babel Machine](https://babel.poltextlab.com). |
|
|
|
## Debugging and issues |
|
This architecture uses the `sentencepiece` tokenizer. In order to run the model before `transformers==4.27` you need to install it manually. |
|
|
|
If you encounter a `RuntimeError` when loading the model using the `from_pretrained()` method, adding `ignore_mismatched_sizes=True` should solve the issue. |
|
|