File size: 2,863 Bytes
d468c58
 
 
 
5d47454
d468c58
55793b6
d468c58
 
 
 
 
5d47454
d468c58
5d47454
 
d468c58
 
 
5d47454
d468c58
5d47454
d468c58
5d47454
 
 
d468c58
5d47454
 
 
 
 
 
 
 
 
55793b6
 
 
 
 
 
 
5d47454
 
55793b6
 
 
 
 
 
d468c58
55793b6
d468c58
55793b6
d468c58
55793b6
 
 
 
 
 
 
 
 
 
 
 
 
d468c58
 
55793b6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
tags:
- generated_from_keras_callback
model-index:
- name: twitter-roberta-base-emotion-multilabel-latest
  results: []
pipeline_tag: text-classification
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# twitter-roberta-base-emotion-multilabel-latest

This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-2021-124m](https://huggingface.co/cardiffnlp/twitter-roberta-base-2021-124m) on the 
[`SemEval 2018 - Task 1 Affect in Tweets`](https://aclanthology.org/S18-1001/) `(subtask: E-c / multilabel classification)`.



## Performance

Following metrics are achieved on the test split:

- F1 (micro): 0.7218
- F1 (macro): 0.5746  
- Jaccard Index (samples): 0.6073: 

### Usage
#### 1. [tweetnlp][https://pypi.org/project/tweetnlp/]
Install tweetnlp via pip.
```shell
pip install tweetnlp
```
Load the model in python.
```python
import tweetnlp

model = tweetnlp.load_model('topic_classification', model_name='cardiffnlp/twitter-roberta-base-emotion-multilabel-latest')

model.predict("I am so happy and sad at the same time")

>> {'label': ['joy', 'sadness']}

```
#### 2. pipeline
```shell
pip install -U tensorflow==2.10
```

```python
from transformers import pipeline

pipe = pipeline("text-classification", model="cardiffnlp/twitter-roberta-base-emotion-multilabel-latest", return_all_scores=True)

pipe("I am so happy and sad at the same time")

>> [[{'label': 'anger', 'score': 0.0059011634439229965},
  {'label': 'anticipation', 'score': 0.024502484127879143},
  {'label': 'disgust', 'score': 0.016748998314142227},
  {'label': 'fear', 'score': 0.20184014737606049},
  {'label': 'joy', 'score': 0.9260002970695496},
  {'label': 'love', 'score': 0.13167349994182587},
  {'label': 'optimism', 'score': 0.32711178064346313},
  {'label': 'pessimism', 'score': 0.08952841907739639},
  {'label': 'sadness', 'score': 0.8542942404747009},
  {'label': 'surprise', 'score': 0.059213291853666306},
  {'label': 'trust', 'score': 0.01618659868836403}]]

```


### Reference 
```
@inproceedings{camacho-collados-etal-2022-tweetnlp,
    title={{T}weet{NLP}: {C}utting-{E}dge {N}atural {L}anguage {P}rocessing for {S}ocial {M}edia},
    author={Camacho-Collados, Jose and Rezaee, Kiamehr and Riahi, Talayeh and Ushio, Asahi and Loureiro, Daniel and Antypas, Dimosthenis and Boisson, Joanne and Espinosa-Anke, Luis and Liu, Fangyu and Mart{\'\i}nez-C{\'a}mara, Eugenio and others},
    author = "Ushio, Asahi  and
      Camacho-Collados, Jose",
    booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations",
    month = nov,
    year = "2022",
    address = "Abu Dhabi, U.A.E.",
    publisher = "Association for Computational Linguistics",
}

```