Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -69,15 +69,27 @@ def gguf_transformers(model_name, sl, tl, input_text):
|
|
| 69 |
|
| 70 |
def gguf_ctransformers(model_name, sl, tl, input_text):
|
| 71 |
from ctransformers import AutoModelForCausalLM
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 72 |
|
| 73 |
-
|
| 74 |
-
filename = "TowerInstruct-7B-v0.1.Q4_K_M.gguf"
|
| 75 |
model = AutoModelForCausalLM.from_pretrained(
|
| 76 |
-
|
| 77 |
model_type="llama"
|
| 78 |
)
|
| 79 |
prompt = f"Translate the following text from {sl} into {tl}.\n{sl}: {input_text}\n{tl}:"
|
| 80 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 81 |
|
| 82 |
def gguf(model_name, sl, tl, input_text):
|
| 83 |
# 1. Pull in tokenizer (handles special tokens)
|
|
@@ -213,7 +225,7 @@ def translate_text(input_text: str, sselected_language: str, tselected_language:
|
|
| 213 |
return translated_text, message_text
|
| 214 |
|
| 215 |
if 'gguf' in model_name.lower():
|
| 216 |
-
translated_text =
|
| 217 |
return translated_text, message_text
|
| 218 |
|
| 219 |
if 'teuken' in model_name.lower():
|
|
|
|
| 69 |
|
| 70 |
def gguf_ctransformers(model_name, sl, tl, input_text):
|
| 71 |
from ctransformers import AutoModelForCausalLM
|
| 72 |
+
from huggingface_hub import hf_hub_download
|
| 73 |
+
|
| 74 |
+
# 1. Fetch the exact .gguf into your local cache
|
| 75 |
+
gguf_path = hf_hub_download(
|
| 76 |
+
repo_id="TheBloke/TowerInstruct-7B-v0.1-GGUF",
|
| 77 |
+
filename="TowerInstruct-7B-v0.1.Q4_K_M.gguf"
|
| 78 |
+
)
|
| 79 |
|
| 80 |
+
# 2. Load once at module scope (so you don’t reload on every call)
|
|
|
|
| 81 |
model = AutoModelForCausalLM.from_pretrained(
|
| 82 |
+
gguf_path,
|
| 83 |
model_type="llama"
|
| 84 |
)
|
| 85 |
prompt = f"Translate the following text from {sl} into {tl}.\n{sl}: {input_text}\n{tl}:"
|
| 86 |
+
out = model.generate(
|
| 87 |
+
prompt,
|
| 88 |
+
max_new_tokens=256,
|
| 89 |
+
temperature=0.2,
|
| 90 |
+
stop=[f"{tl}:"]
|
| 91 |
+
)
|
| 92 |
+
return out[0]
|
| 93 |
|
| 94 |
def gguf(model_name, sl, tl, input_text):
|
| 95 |
# 1. Pull in tokenizer (handles special tokens)
|
|
|
|
| 225 |
return translated_text, message_text
|
| 226 |
|
| 227 |
if 'gguf' in model_name.lower():
|
| 228 |
+
translated_text = gguf_ctransformers(model_name, sselected_language, tselected_language, input_text)
|
| 229 |
return translated_text, message_text
|
| 230 |
|
| 231 |
if 'teuken' in model_name.lower():
|