import gradio as gr from transformers import AutoTokenizer, AutoModelForCausalLM from transformers import pipeline # tokenizer = AutoTokenizer.from_pretrained("cerebras/Cerebras-GPT-13B") # model = AutoModelForCausalLM.from_pretrained("cerebras/Cerebras-GPT-13B") # tokenizer = AutoTokenizer.from_pretrained("cerebras/Cerebras-GPT-2.7B") model = AutoModelForCausalLM.from_pretrained("cerebras/Cerebras-GPT-2.7B") text = "Generative AI is " # pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) pipe = pipeline("text-generation", model=model) def greet(speech): generated_text = pipe(speech, max_length=50, do_sample=False, no_repeat_ngram_size=2)[0] return generated_text['generated_text'] # def greet(speech): # inputs = tokenizer(speech, return_tensors="pt") # outputs = model.generate(**inputs, num_beams=5, # max_new_tokens=50, early_stopping=True, # no_repeat_ngram_size=2) # text_output = tokenizer.batch_decode(outputs, skip_special_tokens=True) # return text_output[0] iface = gr.Interface(fn=greet, inputs="text", outputs="text") iface.launch(share=True)