metadata
license: apache-2.0
base_model: alex-miller/ODABert
tags:
- generated_from_trainer
metrics:
- accuracy
- f1
- precision
- recall
model-index:
- name: iati-climate-multi-classifier-weighted2
results: []
datasets:
- alex-miller/iati-policy-markers
language:
- en
- fr
- es
- de
pipeline_tag: text-classification
widget:
- text: >-
VCA WWF Bolivia The programme will focus on women, young people and
indigenous population living in the transboundary Pantanal - Chaco
ecoregions (PACHA - Paraguay and Bolivia). Its objective is to “amplify
their voices”, to ensure that they are participating, heard and taken into
account in designing solutions for climate transition and common agendas
to reach climate justice.
example_title: Positive
- text: >-
HIV/AIDS prevention by education and awareness raising with emphasis on
gender issues/El Salvador
example_title: Negative
iati-climate-multi-classifier-weighted2
This model is a fine-tuned version of alex-miller/ODABert on a subset of the alex-miller/iati-policy-markers dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7080
- Accuracy: 0.8541
- F1: 0.7121
- Precision: 0.6265
- Recall: 0.8248
Model description
This model has been trained to identify both significant and principal climate mitigation and climate adaptation project titles and/or descriptions.
Intended uses & limitations
As many of the donors in the training dataset have mixed up Adaptation and Mitigation, the model's ability to differentiate the two isn't perfect. But the sigmoid of the model logits do bias toward the correct class.
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-06
- train_batch_size: 24
- eval_batch_size: 24
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
Training results
Training Loss | Epoch | Step | Accuracy | F1 | Validation Loss | Precision | Recall |
---|---|---|---|---|---|---|---|
0.7689 | 1.0 | 1951 | 0.7993 | 0.6421 | 0.6477 | 0.5264 | 0.8230 |
0.6217 | 2.0 | 3902 | 0.8303 | 0.6737 | 0.6269 | 0.5814 | 0.8010 |
0.5834 | 3.0 | 5853 | 0.8266 | 0.6761 | 0.6101 | 0.5715 | 0.8276 |
0.5571 | 4.0 | 7804 | 0.8461 | 0.6933 | 0.6169 | 0.6144 | 0.7954 |
0.5323 | 5.0 | 9755 | 0.8366 | 0.6869 | 0.6050 | 0.5913 | 0.8194 |
0.5126 | 6.0 | 11706 | 0.8327 | 0.6867 | 0.6047 | 0.5815 | 0.8385 |
0.4968 | 7.0 | 13657 | 0.8408 | 0.6938 | 0.6098 | 0.5989 | 0.8244 |
0.4893 | 8.0 | 15608 | 0.6040 | 0.8348 | 0.6895 | 0.5854 | 0.8387 |
0.4702 | 9.0 | 17559 | 0.6342 | 0.8508 | 0.7050 | 0.6211 | 0.8151 |
0.4514 | 10.0 | 19510 | 0.6210 | 0.8383 | 0.6946 | 0.5918 | 0.8404 |
0.4323 | 11.0 | 21461 | 0.6340 | 0.8402 | 0.6991 | 0.5943 | 0.8487 |
0.4193 | 12.0 | 23412 | 0.6407 | 0.8433 | 0.7005 | 0.6020 | 0.8375 |
0.407 | 13.0 | 25363 | 0.6602 | 0.8526 | 0.7094 | 0.6237 | 0.8223 |
0.3944 | 14.0 | 27314 | 0.6588 | 0.8441 | 0.7026 | 0.6029 | 0.8419 |
0.3834 | 15.0 | 29265 | 0.6881 | 0.8529 | 0.7110 | 0.6233 | 0.8274 |
0.3738 | 16.0 | 31216 | 0.7029 | 0.8575 | 0.7146 | 0.6359 | 0.8155 |
0.3686 | 17.0 | 33167 | 0.6929 | 0.8524 | 0.7102 | 0.6224 | 0.8271 |
0.3607 | 18.0 | 35118 | 0.7069 | 0.8545 | 0.7127 | 0.6272 | 0.8253 |
0.3556 | 19.0 | 37069 | 0.7072 | 0.8543 | 0.7118 | 0.6274 | 0.8225 |
0.3523 | 20.0 | 39020 | 0.7080 | 0.8541 | 0.7121 | 0.6265 | 0.8248 |
Framework versions
- Transformers 4.41.0
- Pytorch 2.0.1
- Datasets 2.19.1
- Tokenizers 0.19.1