Edit model card

This classification model is based on DeepPavlov/rubert-base-cased-sentence. The model should be used to produce relevance and specificity of the last message in the context of a dialogue.

The labels explanation:

  • relevance: is the last message in the dialogue relevant in the context of the full dialogue.
  • specificity: is the last message in the dialogue interesting and promotes the continuation of the dialogue.

It is pretrained on a large corpus of dialog data in unsupervised manner: the model is trained to predict whether last response was in a real dialog, or it was pulled from some other dialog at random. Then it was finetuned on manually labelled examples (dataset will be posted soon).

The model was trained with three messages in the context and one response. Each message was tokenized separately with max_length = 32.

The performance of the model on validation split (dataset will be posted soon) (with the best thresholds for validation samples):

threshold f0.5 ROC AUC
relevance 0.49 0.84 0.79
specificity 0.53 0.83 0.83

How to use:

import torch
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained('tinkoff-ai/response-quality-classifier-base')
model = AutoModelForSequenceClassification.from_pretrained('tinkoff-ai/response-quality-classifier-base')
inputs = tokenizer('[CLS]привет[SEP]привет![SEP]как дела?[RESPONSE_TOKEN]норм, у тя как?', max_length=128, add_special_tokens=False, return_tensors='pt')
with torch.inference_mode():
    logits = model(**inputs).logits
    probas = torch.sigmoid(logits)[0].cpu().detach().numpy()
relevance, specificity = probas

The app where you can easily interact with this model.

The work was done during internship at Tinkoff by egoriyaa, mentored by solemn-leader.

Downloads last month
32
Hosted inference API
Text Classification
Examples
Examples
This model can be loaded on the Inference API on-demand.

Space using tinkoff-ai/response-quality-classifier-base