Edit model card

This is a ruBERT-conversational model trained on the mixture of 3 paraphrase detection datasets:

The model can be used to assess semantic similarity of Russian sentences.

Training notebook: task_oriented_TST/similarity/cross_encoders/russian/train_russian_paraphrase_detector__fixed.ipynb (in a private repo).

Training parameters:

  • optimizer: Adam
  • lr=1e-5
  • batch_size=32
  • epochs=3

ROC AUC on the development data:

source         score
detox          0.821665
paraphraser    0.848287
rupaws_qqp     0.761481
rupaws_wiki    0.844093

Pleas see also the documentation of SkolkovoInstitute/ruRoberta-large-paraphrase-v1 that performs better on this task.

Downloads last month
357
Safetensors
Model size
178M params
Tensor type
I64
·
F32
·
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train s-nlp/rubert-base-cased-conversational-paraphrase-v1