| from transformers import AutoModelForCausalLM, AutoTokenizer |
| import torch |
| import gradio as gr |
|
|
| |
| model_name = "Dav66/Te" |
|
|
| |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
| |
| model = AutoModelForCausalLM.from_pretrained( |
| model_name, |
| device_map="auto", |
| offload_folder="./offload", |
| torch_dtype=torch.float16 |
| ) |
|
|
| |
| def generate_text(prompt): |
| with torch.no_grad(): |
| inputs = tokenizer(prompt, return_tensors="pt") |
| outputs = model.generate(**inputs, max_new_tokens=50) |
| text = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| return text |
|
|
| |
| iface = gr.Interface( |
| fn=generate_text, |
| inputs=gr.Textbox(lines=2, placeholder="اكتب هنا النص..."), |
| outputs="text", |
| title="توليد نص بالعربي", |
| description="تطبيق بسيط يعرض موديلك على Hugging Face Space بكفاءة على CPU" |
| ) |
|
|
| iface.launch() |