File size: 782 Bytes
893032f
1
{"valid": {"f1": 37.794928335170894, "recall": 52.28798047589994, "precision": 29.59254143646409, "summary": "              precision    recall  f1-score   support\n\n      entity       0.30      0.52      0.38      3278\n\n   micro avg       0.30      0.52      0.38      3278\n   macro avg       0.30      0.52      0.38      3278\nweighted avg       0.30      0.52      0.38      3278\n"}, "test": {"f1": 38.0477862076775, "recall": 51.69230769230769, "precision": 30.102132234366604, "summary": "              precision    recall  f1-score   support\n\n      entity       0.30      0.52      0.38      3250\n\n   micro avg       0.30      0.52      0.38      3250\n   macro avg       0.30      0.52      0.38      3250\nweighted avg       0.30      0.52      0.38      3250\n"}}