| import gradio as gr |
| import torch |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| import os |
|
|
| |
| |
| MODEL_ID = "TinyLlama/TinyLlama-1.1B-Chat-v1.0" |
|
|
| |
|
|
| |
| def load_model(): |
| """تحميل نموذج Phi-3-Mini المكمم (أقل استهلاكاً للذاكرة).""" |
| try: |
| |
| device = torch.device("cpu") |
| print(f"✅ سيتم تشغيل النموذج على: {device}") |
|
|
| |
| model = AutoModelForCausalLM.from_pretrained( |
| MODEL_ID, |
| torch_dtype=torch.float32, |
| load_in_4bit=True, |
| device_map="auto" |
| ).to(device) |
| |
| tokenizer = AutoTokenizer.from_pretrained(MODEL_ID) |
| |
| return model, tokenizer, device |
| |
| except Exception as e: |
| print(f"❌ فشل تحميل نموذج Phi-3-Mini: {e}") |
| return None, None, None |
|
|
| |
| model, tokenizer, device = load_model() |
|
|
| |
| |