EfektMotyla commited on
Commit
38026de
·
verified ·
1 Parent(s): 2108ce2

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -20
app.py CHANGED
@@ -41,28 +41,25 @@ sentiment_model = AutoModelForSequenceClassification.from_pretrained(
41
  ).to(device)
42
 
43
  # ────────────────────── modele tłumaczeń (on-line) ─────────
44
- pl_to_en_dir = snapshot_download("Helsinki-NLP/opus-mt-pl-en", token=hf_token)
45
- en_to_pl_dir = snapshot_download("gsarti/opus-mt-tc-en-pl", token=hf_token)
46
-
47
- # 📚 Ładowanie tokenizerów i modeli
48
- pl_to_en_tokenizer = MarianTokenizer.from_pretrained(pl_to_en_dir)
49
- pl_to_en_model = MarianMTModel.from_pretrained(pl_to_en_dir).to(device)
50
-
51
- en_to_pl_tokenizer = MarianTokenizer.from_pretrained(en_to_pl_dir)
52
- en_to_pl_model = MarianMTModel.from_pretrained(en_to_pl_dir).to(device)
 
 
 
53
 
54
- # 🔁 Funkcje tłumaczeń
55
- def translate_pl_to_en(texts):
56
- inputs = pl_to_en_tokenizer(texts, return_tensors="pt", padding=True, truncation=True).to(device)
57
- with torch.no_grad():
58
- translated = pl_to_en_model.generate(**inputs)
59
- return pl_to_en_tokenizer.batch_decode(translated, skip_special_tokens=True)
60
 
61
- def translate_en_to_pl(texts):
62
- inputs = en_to_pl_tokenizer(texts, return_tensors="pt", padding=True, truncation=True).to(device)
63
- with torch.no_grad():
64
- translated = en_to_pl_model.generate(**inputs)
65
- return en_to_pl_tokenizer.batch_decode(translated, skip_special_tokens=True)
66
  # ────────────────────── schemy Pydantic ────────────────────
67
  class Comment(BaseModel):
68
  text: str
 
41
  ).to(device)
42
 
43
  # ────────────────────── modele tłumaczeń (on-line) ─────────
44
+ HF_CACHE_DIR = "/tmp/hf_cache"
45
+ os.makedirs(HF_CACHE_DIR, exist_ok=True)
46
+ os.environ["HF_HOME"] = HF_CACHE_DIR
47
+ os.environ["TRANSFORMERS_CACHE"] = HF_CACHE_DIR
48
+
49
+ # Pobieramy modele
50
+ pl_to_en_dir = snapshot_download(
51
+ "Helsinki-NLP/opus-mt-pl-en", token=hf_token, cache_dir=HF_CACHE_DIR
52
+ )
53
+ en_to_pl_dir = snapshot_download(
54
+ "gsarti/opus-mt-tc-en-pl", token=hf_token, cache_dir=HF_CACHE_DIR
55
+ )
56
 
57
+ # Ładujemy
58
+ pl_to_en_tok = MarianTokenizer.from_pretrained(pl_to_en_dir)
59
+ pl_to_en_mod = MarianMTModel.from_pretrained(pl_to_en_dir).to(device)
 
 
 
60
 
61
+ en_to_pl_tok = MarianTokenizer.from_pretrained(en_to_pl_dir)
62
+ en_to_pl_mod = MarianMTModel.from_pretrained(en_to_pl_dir).to(device)
 
 
 
63
  # ────────────────────── schemy Pydantic ────────────────────
64
  class Comment(BaseModel):
65
  text: str