File size: 3,946 Bytes
49910c1
 
 
 
 
 
53d1a3f
 
49910c1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8bf08dd
49910c1
 
cff8dd4
49910c1
b791504
49910c1
8bf08dd
 
49910c1
 
 
f6f6141
 
c6f903a
 
f6f6141
 
 
 
 
 
 
49910c1
 
 
 
 
 
 
 
 
 
 
 
 
 
f6790c1
49910c1
f6790c1
49910c1
 
 
 
f6790c1
49910c1
 
8bf08dd
53d1a3f
8bf08dd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9e5fc6e
9e9e2c0
 
 
 
 
 
 
 
 
 
 
 
9e5fc6e
49910c1
 
 
 
 
af584d6
 
 
 
d271102
af584d6
 
 
 
 
 
 
 
 
 
 
ddc2f74
 
af584d6
 
 
 
 
 
 
d271102
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
---
license: apache-2.0
tags:
- 
datasets:
- EXIST Dataset
- MeTwo Machismo and Sexism Twitter Identification dataset

metrics:
- accuracy
model-index:
- name: twitter_sexismo-finetuned-exist2021
  results:
  - task:
      name: Text Classification
      type: text-classification
    dataset:
      name: EXIST Dataset
      type: EXIST Dataset
      args: es
    metrics:
    - name: Accuracy
      type: accuracy
      value: 0.83
---

# twitter_sexismo-finetuned-exist2021

This model is a fine-tuned version of [pysentimiento/robertuito-hate-speech](https://huggingface.co/pysentimiento/robertuito-hate-speech) on the EXIST dataset and MeTwo: Machismo and Sexism Twitter Identification dataset https://github.com/franciscorodriguez92/MeTwo.
It achieves the following results on the evaluation set:
- Loss: 0.54
- Accuracy: 0.83

## Model description

Modelo para el Hackaton de Somos NLP para detección de sexismo en twitts en español. Creado por:

medardodt

MariaIsabel

ManRo

lucel172

robertou2

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- my_learning_rate = 5E-5 
- my_adam_epsilon = 1E-8 
- my_number_of_epochs = 8
- my_warmup = 3
- my_mini_batch_size = 32
- optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 8

### Training results
Epoch 	Training Loss 	Validation Loss 	Accuracy 	F1 	Precision 	Recall

1 	0.389900 	0.397857 	0.827133 	0.699620 	0.786325 	0.630137

2 	0.064400 	0.544625 	0.831510 	0.707224 	0.794872 	0.636986

3 	0.004800 	0.837723 	0.818381 	0.704626 	0.733333 	0.678082

4 	0.000500 	1.045066 	0.820569 	0.702899 	0.746154 	0.664384

5 	0.000200 	1.172727 	0.805252 	0.669145 	0.731707 	0.616438

6 	0.000200 	1.202422 	0.827133 	0.720848 	0.744526 	0.698630

7 	0.000000 	1.195012 	0.827133 	0.718861 	0.748148 	0.691781

8 	0.000100 	1.215515 	0.824945 	0.705882 	0.761905 	0.657534

9 	0.000100 	1.233099 	0.827133 	0.710623 	0.763780 	0.664384

10 	0.000100 	1.237268 	0.829322 	0.713235 	0.769841 	0.664384

|Epoch|Training Loss|Validation Loss|Accuracy|F1|Precision|Precision|
|----|-------|-------|-------|-------|-------|-------| 
|1|0.389900 |0.397857 |0.827133 |0.699620 |0.786325 |0.630137 |
|2|0.064400 |0.544625 |0.831510 |0.707224 |0.794872 |0.636986 | 
|3|0.004800 |0.837723 |0.818381 |0.704626 |0.733333 |0.678082 | 
|4|0.000500 |1.045066 |0.820569 | 0.702899 |0.746154 |0.664384 |
|5|0.000200 |1.172727 |0.805252 |0.669145 |0.731707 |0.616438 | 
|6|0.000200 |1.202422 |0.827133 |0.720848 |0.744526 |0.698630 | 
|7|0.000000 |1.195012 |0.827133 |0.718861 |0.748148 |0.691781 | 
|8|0.000100 |1.215515 |0.824945 |0.705882 |0.761905 |0.657534 | 
|9|0.000100|1.233099 |0.827133 |0.710623 |0.763780 |0.664384 | 
|10|0.000100|1.237268 |0.829322 |0.713235 |0.769841 |0.664384 | 

### Framework versions

- Transformers 4.17.0
- Pytorch 1.10.0+cu111
- Tokenizers 0.11.6


## Model in Action
Fast usage with pipelines:
``` python
###libraries required
!pip install transformers
from transformers import pipeline

### usage pipelines
model_checkpoint = "hackathon-pln-es/twitter_sexismo-finetuned-exist2021-metwo" 
pipeline_nlp = pipeline("text-classification", model=model_checkpoint)
pipeline_nlp("mujer al volante peligro!") 
#pipeline_nlp("¡me encanta el ipad!") 
#pipeline_nlp (["mujer al volante peligro!", "Los hombre tienen más manias que las mujeres", "me encanta el ipad!"] )

# OUTPUT MODEL #
# LABEL_0: "NON SEXISM"or LABEL_1: "SEXISM"  and score: probability of accuracy per model.

# [{'label': 'LABEL_1', 'score': 0.9967633485794067}]
# [{'label': 'LABEL_0', 'score': 0.9934417009353638}]

#[{‘label': 'LABEL_1', 'score': 0.9967633485794067},
# {'label': 'LABEL_1', 'score': 0.9755664467811584},
# {'label': 'LABEL_0', 'score': 0.9955045580863953}]
```