TiberiuCristianLeon commited on
Commit
f7c3472
·
verified ·
1 Parent(s): e33a1bb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -5
app.py CHANGED
@@ -69,15 +69,27 @@ def gguf_transformers(model_name, sl, tl, input_text):
69
 
70
  def gguf_ctransformers(model_name, sl, tl, input_text):
71
  from ctransformers import AutoModelForCausalLM
 
 
 
 
 
 
 
72
 
73
- repo_id = "TheBloke/TowerInstruct-7B-v0.1-GGUF"
74
- filename = "TowerInstruct-7B-v0.1.Q4_K_M.gguf"
75
  model = AutoModelForCausalLM.from_pretrained(
76
- "https://huggingface.co/TheBloke/TowerInstruct-7B-v0.1-GGUF/resolve/main/towerinstruct-7b-v0.1.Q4_K_M.gguf",
77
  model_type="llama"
78
  )
79
  prompt = f"Translate the following text from {sl} into {tl}.\n{sl}: {input_text}\n{tl}:"
80
- return model.generate(prompt)[0]
 
 
 
 
 
 
81
 
82
  def gguf(model_name, sl, tl, input_text):
83
  # 1. Pull in tokenizer (handles special tokens)
@@ -213,7 +225,7 @@ def translate_text(input_text: str, sselected_language: str, tselected_language:
213
  return translated_text, message_text
214
 
215
  if 'gguf' in model_name.lower():
216
- translated_text = gguf_transformers(model_name, sselected_language, tselected_language, input_text)
217
  return translated_text, message_text
218
 
219
  if 'teuken' in model_name.lower():
 
69
 
70
  def gguf_ctransformers(model_name, sl, tl, input_text):
71
  from ctransformers import AutoModelForCausalLM
72
+ from huggingface_hub import hf_hub_download
73
+
74
+ # 1. Fetch the exact .gguf into your local cache
75
+ gguf_path = hf_hub_download(
76
+ repo_id="TheBloke/TowerInstruct-7B-v0.1-GGUF",
77
+ filename="TowerInstruct-7B-v0.1.Q4_K_M.gguf"
78
+ )
79
 
80
+ # 2. Load once at module scope (so you don’t reload on every call)
 
81
  model = AutoModelForCausalLM.from_pretrained(
82
+ gguf_path,
83
  model_type="llama"
84
  )
85
  prompt = f"Translate the following text from {sl} into {tl}.\n{sl}: {input_text}\n{tl}:"
86
+ out = model.generate(
87
+ prompt,
88
+ max_new_tokens=256,
89
+ temperature=0.2,
90
+ stop=[f"{tl}:"]
91
+ )
92
+ return out[0]
93
 
94
  def gguf(model_name, sl, tl, input_text):
95
  # 1. Pull in tokenizer (handles special tokens)
 
225
  return translated_text, message_text
226
 
227
  if 'gguf' in model_name.lower():
228
+ translated_text = gguf_ctransformers(model_name, sselected_language, tselected_language, input_text)
229
  return translated_text, message_text
230
 
231
  if 'teuken' in model_name.lower():