added f1 score
Browse files
README.md
CHANGED
@@ -19,7 +19,11 @@ This model is trained to predict whether two given messages from some group chat
|
|
19 |
# Training details
|
20 |
|
21 |
It's based on [Conversational RuBERT](https://docs.deeppavlov.ai/en/master/features/models/bert.html) (cased, 12-layer, 768-hidden, 12-heads, 180M parameters) that was trained on several social media datasets. We fine-tuned it with the data from several Telegram chats. The positive `reply_to` examples were obtained by natural user annotation. The negative ones were obtained by shuffling the messages.
|
22 |
-
The task perfectly aligns with the Next Sentence Prediction task, so the fine-tuning was done in that manner.
|
|
|
|
|
|
|
|
|
23 |
|
24 |
# Usage
|
25 |
|
|
|
19 |
# Training details
|
20 |
|
21 |
It's based on [Conversational RuBERT](https://docs.deeppavlov.ai/en/master/features/models/bert.html) (cased, 12-layer, 768-hidden, 12-heads, 180M parameters) that was trained on several social media datasets. We fine-tuned it with the data from several Telegram chats. The positive `reply_to` examples were obtained by natural user annotation. The negative ones were obtained by shuffling the messages.
|
22 |
+
The task perfectly aligns with the Next Sentence Prediction task, so the fine-tuning was done in that manner.
|
23 |
+
|
24 |
+
It achieves the 0.83 F1 score on the gold test set from our reply recovery dataset.
|
25 |
+
|
26 |
+
See the [paper](https://www.dialog-21.ru/media/5871/buyanoviplusetal046.pdf) for more details.
|
27 |
|
28 |
# Usage
|
29 |
|