reword / app.py
Grady Harwood
updated model
2abe29a
raw
history blame contribute delete
No virus
1.8 kB
from transformers import pipeline
# from transformers import T5Tokenizer, T5ForConditionalGeneration
import gradio as gr
def pipe(input_text):
# tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-base")
# model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-base")
# input_text = "reword for clarity" + input_text
# input_ids = tokenizer(input_text, return_tensors="pt").input_ids
# outputs = model.generate(input_ids)
# return tokenizer.decode(outputs[0])
# Use a pipeline as a high-level helper
model = pipeline(
task='question-answering',
model="mistralai/Mistral-7B-Instruct-v0.3",
)
output = model(
question="reword for clarity",
context=input_text,
)
return output["answer"]
demo = gr.Interface(
fn=pipe,
inputs=gr.Textbox(lines=7),
outputs="text",
)
demo.launch()
# # pip install -q transformers
# from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
# checkpoint = "CohereForAI/aya-101"
# tokenizer = AutoTokenizer.from_pretrained(checkpoint)
# aya_model = AutoModelForSeq2SeqLM.from_pretrained(checkpoint)
# def generator(input_text):
# inputs = tokenizer.encode("Translate to English: " + input_text, return_tensors="pt")
# outputs = aya_model.generate(inputs, max_new_tokens=128)
# return tokenizer.decode(outputs[0])
# # # Turkish to English translation
# # tur_inputs = tokenizer.encode("Translate to English: Aya cok dilli bir dil modelidir.", return_tensors="pt")
# # tur_outputs = aya_model.generate(tur_inputs, max_new_tokens=128)
# # print(tokenizer.decode(tur_outputs[0]))
# # # Aya is a multi-lingual language model
# demo = gr.Interface(
# fn=generator,
# inputs=gr.Textbox(lines=7),
# outputs="text",
# )
# demo.launch()