Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)


Please try this ➤➤➤ Colab Notebook Demo (click me!)

Context Response width score
I love NLP! Can anyone recommend a nice review paper? 0.701
I love NLP! Me too! 0.029

The width score predicts how likely the response is getting replied.


Dialog Ranking Pretrained Transformers

How likely a dialog response is upvoted 👍 and/or gets replied 💬?

This is what DialogRPT is learned to predict. It is a set of dialog response ranking models proposed by Microsoft Research NLP Group trained on 100 + millions of human feedback data. It can be used to improve existing dialog generation model (e.g., DialoGPT) by re-ranking the generated response candidates.

Quick Links:

We considered the following tasks and provided corresponding pretrained models.

Task Description Pretrained model
Human feedback given a context and its two human responses, predict...
updown ... which gets more upvotes? model card
width ... which gets more direct replies? this model
depth ... which gets longer follow-up thread? model card
Human-like (human vs fake) given a context and one human response, distinguish it with...
human_vs_rand ... a random human response model card
human_vs_machine ... a machine generated response model card


Please create an issue on our repo


    title={Dialogue Response RankingTraining with Large-Scale Human Feedback Data},
    author={Xiang Gao and Yizhe Zhang and Michel Galley and Chris Brockett and Bill Dolan},
Downloads last month