Back to all models
Model card Files and versions Use in transformers
text-classification mask_token: [MASK]
Query this model
πŸ”₯ This model is currently loaded and running on the Inference API. ⚠️ This model could not be loaded by the inference API. ⚠️ This model can be loaded on the Inference API on-demand.
JSON Output
API endpoint  

⚑️ Upgrade your account to access the Inference API

Share Copied link to clipboard

Contributed by

nateraw Nate Raw
3 models


Model description

bert-base-uncased finetuned on the emotion dataset using PyTorch Lightning. Sequence length 128, learning rate 2e-5, batch size 32, 2 GPUs, 4 epochs.

For more details, please see, the emotion dataset on nlp viewer.

Limitations and bias

  • Not the best model, but it works in a pinch I guess...
  • Code not available as I just hacked this together.
  • Follow me on github to get notified when code is made available.

Training data

Data came from HuggingFace's datasets package. The data can be viewed on nlp viewer.

Training procedure


Eval results

val_acc - 0.931 (useless, as this should be precision/recall/f1)

The score was calculated using PyTorch Lightning metrics.