Spaces:
Sleeping
Sleeping
import gradio as gr | |
from transformers import pipeline | |
import os | |
access_token = os.getenv("Token") | |
pipe = pipeline("text-generation", model="meta-llama/Llama-3.2-1B-Instruct", token=access_token) | |
def predict(query): | |
messages = [] | |
messages.append({"role": "system","content": "you are a helpful assistant"}) | |
messages.append({"role": "user","content": query}) | |
return pipe(messages,max_new_tokens=256)[0]["generated_text"][-1]['content'] | |
gradio_app = gr.Interface( | |
predict, | |
inputs="text", outputs="text", | |
title="Llama-3.2-1B-Instruct", | |
) | |
if __name__ == "__main__": | |
gradio_app.launch() |