Spaces:
Sleeping
Sleeping
File size: 998 Bytes
eb91edf c747562 eb91edf cb2adb5 eb91edf cb2adb5 c747562 eb91edf c747562 eb91edf c747562 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
import streamlit as st
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
from model.funcs import execution_time
@st.cache_data
def load_model():
model_path = "17/"
model_name = "sberbank-ai/rugpt3small_based_on_gpt2"
tokenizer = GPT2Tokenizer.from_pretrained(model_name)
model = GPT2LMHeadModel.from_pretrained(model_path)
return tokenizer, model
tokenizer, model = load_model()
@execution_time
def generate_text(promt):
promt = tokenizer.encode(promt, return_tensors="pt")
model.eval()
with torch.no_grad():
out = model.generate(
promt,
do_sample=True,
num_beams=2,
temperature=1.5,
top_p=0.9,
max_length=150,
)
out = list(map(tokenizer.decode, out))[0]
return out
promt = st.text_input("Ask a question")
generate = st.button("Generate")
if generate:
if not promt:
st.write("42")
else:
st.write(generate_text(promt))
|