import streamlit as st import transformers import tensorflow from transformers import AutoTokenizer, TFAutoModelForSeq2SeqLM model_checkpoint = "Modfiededition/t5-base-fine-tuned-on-jfleg" tokenizer = AutoTokenizer.from_pretrained(model_checkpoint) @st.cache(allow_output_mutation=True, suppress_st_warning=True) def load_model(): return pipeline("text2text- generation", model=model_checkpoint) model = load_model() def infer(input_ids): output_sequences = model.generate(inputs["input_ids"]).numpy()[0][1:-1] return output_sequences #prompts st.title("Writing Assistant for you 🦄") textbox = st.text_area('Write your text:', '', height=200, max_chars=1000) #inputs = tokenizer("Grammar: "+sent,return_tensors="tf") #output_sequences = infer(inputs) #generated_sequences = tokenizer.decode(output_ids) #st.write(generated_sequences)