Edit model card

WRIME-fine-tuned BERT base Japanese

This model is a Japanese BERTBASE fine-tuned on the WRIME dataset. It was trained as part of the paper "Emotion Analysis of Writers and Readers of Japanese Tweets on Vaccinations". Fine-tuning code is available at this repo.

Intended uses and limitations

This model can be used to predict intensities scores for eight emotions for writers and readers. Please refer to the Fine-tuning data section for the list of emotions.

Because of the regression fine-tuning task, it is possible for the model to infer scores outside of the range of the scores of the fine-tuning data (score < 0 or score > 4).

Model Architecture, Tokenization, and Pretraining

The Japanese BERTBASE fine-tuned was cl-tohoku/bert-base-japanese-v2. Please refer to their model card for details regarding the model architecture, tokenization, pretraining data, and pretraining procedure.

Fine-tuning data

The model is fine-tuned on WRIME, a dataset of Japanese Tweets annotated with writer and reader emotion intensities. We use version 1 of the dataset. Each Tweet is accompanied by a set of writer emotion intensities (from the author of the Tweet) and three sets of reader emotions (from three annotators). The emotions follow Plutchhik's emotions, namely:

  • joy
  • sadness
  • anticipation
  • surprise
  • anger
  • fear
  • disgust
  • trust

These emotion intensities follow a four-point scale:

emotion intensity emotion presence
0 no
1 weak
2 medium
3 strong

Fine-tuning

The BERT is fine-tuned to directly regress the emotion intensities of the writer and the averaged emotions of the readers from each Tweet, meaning there are 16 outputs (8 emotions per writer/reader).

The fine-tuning was inspired by common BERT fine-tuning procedures. The BERT was fine-tuned on WRIME for 3 epochs using the AdamW optimizer with a learning rate of 2e-5, β1=0.9, β2=0.999, weight decay of 0.01, linear decay, a warmup ratio of 0.01, and a batch size of 32. Training was conducted with an NVIDIA Tesla K80 and finished in 3 hours.

Evaluation results

Below are the MSEs of the BERT on the test split of WRIME.

Annotator Joy Sadness Anticipation Surprise Anger Fear Disgust Trust Overall
Writer 0.658 0.688 0.746 0.542 0.486 0.462 0.664 0.400 0.581
Reader 0.192 0.178 0.211 0.139 0.032 0.147 0.123 0.029 0.131
Both 0.425 0.433 0.479 0.341 0.259 0.304 0.394 0.214 0.356
Downloads last month
50
Safetensors
Model size
111M params
Tensor type
I64
·
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.