Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -148,7 +148,8 @@ theme = gr.themes.Monochrome(
|
|
148 |
|
149 |
def stream(model, code, generate_kwargs):
|
150 |
input_ids = tokenizer(code, return_tensors="pt").to(device)
|
151 |
-
generated_ids = model.generate(**input_ids, **generate_kwargs)
|
|
|
152 |
return tokenizer.decode(generated_ids[0][input_ids["input_ids"].shape[1]:], skip_special_tokens=True).strip()
|
153 |
|
154 |
@spaces.GPU
|
@@ -170,23 +171,22 @@ def generate(
|
|
170 |
seed=42,
|
171 |
)
|
172 |
|
173 |
-
|
174 |
-
|
175 |
-
|
176 |
-
|
177 |
-
|
178 |
-
|
179 |
-
|
180 |
-
|
181 |
-
|
182 |
-
|
183 |
-
|
184 |
-
|
185 |
-
|
186 |
-
|
187 |
-
|
188 |
-
|
189 |
-
output = ""
|
190 |
return output
|
191 |
|
192 |
|
|
|
148 |
|
149 |
def stream(model, code, generate_kwargs):
|
150 |
input_ids = tokenizer(code, return_tensors="pt").to(device)
|
151 |
+
# generated_ids = model.generate(**input_ids, **generate_kwargs)
|
152 |
+
generated_ids = model.generate(**input_ids)
|
153 |
return tokenizer.decode(generated_ids[0][input_ids["input_ids"].shape[1]:], skip_special_tokens=True).strip()
|
154 |
|
155 |
@spaces.GPU
|
|
|
171 |
seed=42,
|
172 |
)
|
173 |
|
174 |
+
if method == "Base":
|
175 |
+
output = stream(basemodel, prompt, generate_kwargs)
|
176 |
+
elif method == "Prefix":
|
177 |
+
output = stream(model_map[library + " Prefix"], prompt, generate_kwargs)
|
178 |
+
elif method == "Evo Prefix" and library in ["SQLModel", "SfePy", "MegEngine"]:
|
179 |
+
output = stream(model_map["Main Evo Prefix"], prompt, generate_kwargs)
|
180 |
+
elif method == "FFT" and library in ["SQLModel", "SfePy", "MegEngine"]:
|
181 |
+
output = stream(model_map[library + " FFT"], prompt, generate_kwargs)
|
182 |
+
elif method == "Evo FFT" and library in ["SQLModel", "SfePy", "MegEngine"]:
|
183 |
+
output = stream(model_map["Main Evo FFT"], prompt, generate_kwargs)
|
184 |
+
elif method == "Full Data FFT" and library in ["SQLModel", "SfePy", "MegEngine"]:
|
185 |
+
output = stream(model_map["Main FD FFT"], prompt, generate_kwargs)
|
186 |
+
elif method == "Evo Prefix" and library in ["LangChain", "LlamaIndex", "DSPy"]:
|
187 |
+
output = stream(model_map["CS Evo Prefix"], prompt, generate_kwargs)
|
188 |
+
else:
|
189 |
+
output = ""
|
|
|
190 |
return output
|
191 |
|
192 |
|