Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,7 +1,6 @@
|
|
| 1 |
import os
|
| 2 |
import gradio as gr
|
| 3 |
import spaces
|
| 4 |
-
import torch
|
| 5 |
|
| 6 |
# 支持的语言选项
|
| 7 |
LANGUAGES = {
|
|
@@ -100,41 +99,57 @@ LANGUAGES = {
|
|
| 100 |
"Western Persian": "Western Persian"
|
| 101 |
}
|
| 102 |
|
| 103 |
-
from
|
| 104 |
-
from vllm import LLM
|
| 105 |
-
from vllm.sampling_params import BeamSearchParams
|
| 106 |
|
| 107 |
-
save_dir = "./model_weights"
|
| 108 |
-
repo_id = "ByteDance-Seed/Seed-X-PPO-7B"
|
| 109 |
-
cache_dir = save_dir + "/cache"
|
| 110 |
|
| 111 |
-
snapshot_download
|
| 112 |
-
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
local_dir_use_symlinks=False,
|
| 116 |
-
resume_download=True,
|
| 117 |
-
allow_patterns=["*.json", "*.safetensors", "*.bin", "*.py", "*.md", "*.txt"],
|
| 118 |
-
)
|
| 119 |
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
|
| 125 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 126 |
|
| 127 |
-
decoding_params = BeamSearchParams(beam_width=4,
|
| 128 |
-
max_tokens=512)
|
| 129 |
|
| 130 |
@spaces.GPU(duration=120)
|
| 131 |
def translate_text(text, source_lang, target_lang):
|
| 132 |
if not text.strip():
|
| 133 |
return "请输入要翻译的文本"
|
| 134 |
try:
|
| 135 |
-
|
| 136 |
-
|
| 137 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
except Exception as e:
|
| 139 |
yield f"翻译出错: {str(e)}"
|
| 140 |
|
|
|
|
| 1 |
import os
|
| 2 |
import gradio as gr
|
| 3 |
import spaces
|
|
|
|
| 4 |
|
| 5 |
# 支持的语言选项
|
| 6 |
LANGUAGES = {
|
|
|
|
| 99 |
"Western Persian": "Western Persian"
|
| 100 |
}
|
| 101 |
|
| 102 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
|
|
|
| 103 |
|
|
|
|
|
|
|
|
|
|
| 104 |
|
| 105 |
+
# from huggingface_hub import snapshot_download
|
| 106 |
+
# save_dir = "./model_weights"
|
| 107 |
+
# repo_id = "ByteDance-Seed/Seed-X-PPO-7B"
|
| 108 |
+
# cache_dir = save_dir + "/cache"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
|
| 110 |
+
# snapshot_download(
|
| 111 |
+
# cache_dir=cache_dir,
|
| 112 |
+
# local_dir=save_dir,
|
| 113 |
+
# repo_id=repo_id,
|
| 114 |
+
# local_dir_use_symlinks=False,
|
| 115 |
+
# resume_download=True,
|
| 116 |
+
# allow_patterns=["*.json", "*.safetensors", "*.bin", "*.py", "*.md", "*.txt"],
|
| 117 |
+
# )
|
| 118 |
+
|
| 119 |
+
device = "cuda"
|
| 120 |
+
MODEL_NAME = "ByteDance-Seed/Seed-X-PPO-7B"
|
| 121 |
+
print("Start dowload")
|
| 122 |
+
def load_model():
|
| 123 |
+
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME,torch_dtype="bfloat16").to(device)
|
| 124 |
+
print(f"Model loaded in {device}")
|
| 125 |
+
return model
|
| 126 |
+
|
| 127 |
+
model = load_model()
|
| 128 |
+
print("Ednd dowload")
|
| 129 |
+
# Loading the tokenizer once, because re-loading it takes about 1.5 seconds each time
|
| 130 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 131 |
|
|
|
|
|
|
|
| 132 |
|
| 133 |
@spaces.GPU(duration=120)
|
| 134 |
def translate_text(text, source_lang, target_lang):
|
| 135 |
if not text.strip():
|
| 136 |
return "请输入要翻译的文本"
|
| 137 |
try:
|
| 138 |
+
input_tokens = (
|
| 139 |
+
tokenizer(text, return_tensors="pt")
|
| 140 |
+
.input_ids[0]
|
| 141 |
+
.cpu()
|
| 142 |
+
.numpy()
|
| 143 |
+
.tolist()
|
| 144 |
+
)
|
| 145 |
+
translated_chunk = model.generate(
|
| 146 |
+
input_ids=torch.tensor([input_tokens]).to(device),
|
| 147 |
+
max_length=len(input_tokens) + 2048,
|
| 148 |
+
num_return_sequences=1,
|
| 149 |
+
)
|
| 150 |
+
full_output = tokenizer.decode(translated_chunk[0], skip_special_tokens=True)
|
| 151 |
+
full_output = full_output.replace(input_text,"")
|
| 152 |
+
yield full_output
|
| 153 |
except Exception as e:
|
| 154 |
yield f"翻译出错: {str(e)}"
|
| 155 |
|