jpcorb20's picture
Update README.md
8873e98

Distilroberta for toxic comment detection

See my GitHub repo toxic-comment-server

The model was trained from DistilRoberta on Kaggle Toxic Comments with the BCEWithLogits loss for Multi-Label prediction. Thus, please use the sigmoid activation on the logits (not made to use the softmax output, e.g. like the HF widget).

Evaluation

F1 scores:

  toxic: 0.72
  severe_toxic: 0.38
  obscene: 0.72
  threat: 0.52
  insult: 0.69
  identity_hate: 0.60

  Macro-F1: 0.61