import torch from transformers import AutoModelForCausalLM, AutoTokenizer import gradio as gr # torch.set_default_device("cuda") model = AutoModelForCausalLM.from_pretrained("microsoft/phi-2", torch_dtype=torch.float32, device_map="cpu", trust_remote_code=True) tokenizer = AutoTokenizer.from_pretrained("microsoft/phi-2", trust_remote_code=True) def greet(text): inputs = tokenizer('Instruct: '+text+'\nOutput: ', return_tensors="pt", return_attention_mask=False) outputs = model.generate(**inputs, max_length=200) _text = tokenizer.batch_decode(outputs)[0] return _text demo = gr.Interface(fn=greet, inputs="text", outputs="text") demo.launch()