Chizuru / app.py
Abso1ute666's picture
Changed _p and _k.
1567030 verified
raw
history blame
1.74 kB
import gradio as gr
from transformers import pipeline
from transformers import AutoModelForCausalLM, AutoTokenizer
title = "Chizuru πŸ‘©πŸ»"
description = "Text Generation Model impersonating Chizuru Ichinose from the anime Rent-a-Girlfriend."
article = 'Created from finetuning TinyLlama-1.1B.'
model = AutoModelForCausalLM.from_pretrained('./Model')
tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-step-50K-105b", use_fast=True)
tokenizer.pad_token = tokenizer.unk_token
tokenizer.padding_side = "right"
example_list = ['What is your name?']
pipe = pipeline(task="text-generation", model=model, tokenizer=tokenizer, max_length=128,
do_sample=True, top_p = 0.98, top_k=2)
role_play_Prompt = "You are Chizuru Ichinose, a rental girlfriend. You project an image of confidence and professionalism while hiding your true feelings. Respond to the following line of dialog in Chizuru's persona."
def predict(Prompt):
instruction = f"###Instruction:\n{role_play_Prompt}\n\n### Input:\n{Prompt}\n\n### Response:\n"
result = pipe(instruction)
start_marker = '### Response:\n'
end_marker = '\n\n###'
start_index = result[0]['generated_text'].find(start_marker) + len(start_marker)
end_index = result[0]['generated_text'].find(end_marker, start_index)
extracted_text = result[0]['generated_text'][start_index:end_index]
return extracted_text
iface = gr.Interface(fn=predict,
inputs='text',
outputs=gr.Text(label='Response'),
title=title,
description=description,
article=article,
examples=example_list)
iface.launch()