Sean Halpin
Update README.md
bd850cf
metadata
thumbnail: https://huggingface.co/front/thumbnails/dialogpt.png
tags:
  - conversational
license: mit

dialogGPT-homer-simpson

This model has been fine tuned with the entire scripts of Homer Simpson from the T.V. show The Simpsons It will give some nice answers seemingly from Homers brain in the Simpsons Universe during single turn conversation, letting you chat to Homer Simpson

A State-of-the-Art Large-scale Pretrained Response generation model (DialoGPT)

DialoGPT is a SOTA large-scale pretrained dialogue response generation model for multiturn conversations. The human evaluation results indicate that the response generated from DialoGPT is comparable to human response quality under a single-turn conversation Turing test. The model is trained on 147M multi-turn dialogue from Reddit discussion thread.

  • Multi-turn generation examples from an interactive environment:
Role Response
User Who are you?
HomerBot Homer Simpson .
User What is your favorite Restaurant ?
HomerBot Moes Tavern.
User Have you ever been in a band?!
HomerBot no.

Please find the information about preprocessing, training and full details of the DialoGPT in the original DialoGPT repository

ArXiv paper: https://arxiv.org/abs/1911.00536

How to use Multi-Turn

NOTE: Multi-Turn seems to be broken, after a few exchanges the output will mostly be exclamation marks.

Now we are ready to try out how the model works as a chatting partner!

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch


tokenizer = AutoTokenizer.from_pretrained("shalpin87/dialoGPT-homer-simpson")
model = AutoModelForCausalLM.from_pretrained("shalpin87/dialoGPT-homer-simpson")

# Let's chat for 5 lines
for step in range(5):
    # encode the new user input, add the eos_token and return a tensor in Pytorch
    new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt')

    # append the new user input tokens to the chat history
    bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids

    # generated a response while limiting the total chat history to 1000 tokens, 
    chat_history_ids = model.generate(bot_input_ids, max_length=1000, pad_token_id=tokenizer.eos_token_id)

    # pretty print last ouput tokens from bot
    print("DialoG-PT-HomerBot: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True)))

How to use Single Turn

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch


tokenizer = AutoTokenizer.from_pretrained("shalpin87/dialoGPT-homer-simpson")
model = AutoModelForCausalLM.from_pretrained("shalpin87/dialoGPT-homer-simpson")

questions = [
    "What is your name?",
    "Who are you?",
    "Where do you work?",
    "Who really killed Mr Burns?",
    "Have you ever stolen from the Kwik-E-Mart?",
    "Did you kill Frank Grimes?",
    "Who was the worst member of the Be Sharps?",
    "Hey where did Barney go?",
    "What is your favorite bar to have a beer?",
    "What is the best beer in Springfield?",
    "Is Bart working for the Mob?",
    "I think there was an incident in sector 7 G",
    "Is Ned Flanders house okay?",
    "Oh my god it's Sideshow Bob",
    "What is a Flaming Moe?",
    "What is happening to Apu?",
    "Who quit the band?"
]

# Let's chat
for step in range(len(questions)):
    print("***************************************")
    print("Q. {}".format(questions[step]))
    new_user_input_ids = tokenizer.encode(questions[step] + tokenizer.eos_token, return_tensors='pt')
    # new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt')

    bot_input_ids = new_user_input_ids

    num_return_seqs=1

    chat_history_ids = model.generate(
        bot_input_ids, 
        max_length=200,
        pad_token_id=tokenizer.eos_token_id,  
        no_repeat_ngram_size=3,       
        do_sample=True, 
        top_k=200, 
        top_p=0.55,
        temperature = 0.85,
        num_return_sequences=num_return_seqs
    )
    
    botname = "HomerBot"
    for i in range(0,num_return_seqs):
      print("{}:{}: {}".format(i, botname, tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][i], skip_special_tokens=True)))

    chat_history_ids = []

Sample Output

***************************************
Q. What is your name?
0:HomerBot: Homer Simpson.
***************************************
Q. Who are you?
0:HomerBot: Homer Simpson.
***************************************
Q. Where do you work?
0:HomerBot: I work at a nuclear power plant.
***************************************
Q. Who really killed Mr Burns?
0:HomerBot: I dunno. Some guy named Ot-to.
***************************************
Q. Have you ever stolen from the Kwik-E-Mart?
0:HomerBot: No.
***************************************
Q. Did you kill Frank Grimes?
0:HomerBot: No.
***************************************
Q. Who was the worst member of the Be Sharps?
0:HomerBot: Homer Simpson, sir.
***************************************
Q. Hey where did Barney go?
0:HomerBot: He's been eating that thing for a week.
***************************************
Q. What is your favorite bar to have a beer?
0:HomerBot: The Flanders' Club.
***************************************
Q. What is the best beer in Springfield?
0:HomerBot: Pepi!
***************************************
Q. Is Bart working for the Mob?
0:HomerBot: I think he's trying to tell us something.
***************************************
Q. I think there was an incident in sector 7 G
0:HomerBot: Uh, let's not jump to any conclusions. The dog didn't run away.