File size: 1,265 Bytes
2ce1fb9
 
 
9b7d8b6
 
 
6cb6276
2ce1fb9
 
29aa142
9b7d8b6
53eb945
 
29aa142
53eb945
f383b0c
 
29aa142
 
 
 
 
 
 
 
53eb945
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
language: en
tags:
- emotion-classification
datasets:
- go-emotions
- bdotloh/empathetic-dialogues-contexts
---


# Model Description
Yet another Transformer model fine-tuned for approximating another non-linear mapping between X and Y? That's right!
This is your good ol' emotion classifier - given an input text, the model outputs a probability distribution over a set of pre-selected emotion words. In this case, it is 32, which is the number of emotion classes in the [Empathetic Dialogues](https://huggingface.co/datasets/bdotloh/empathetic-dialogues-contexts) dataset.

This model is built "on top of" a [distilbert-base-uncased model fine-tuned on the go-emotions dataset](https://huggingface.co/bhadresh-savani/bert-base-go-emotion). Y'all should really check out that model, it even contains a jupyter notebook file that illustrates how the model was trained (bhadresh-savani if you see this, thank you!).  


## Training data

## Training procedure

### Preprocessing

## Evaluation results

### Limitations and bias
Well where should we begin...

EmpatheticDialogues: 
1) Unable to ascertain the degree of cultural specificity for the context that a respondent described when given an emotion label (i.e., p(description | emotion, *culture*))
2) ...