Spaces:
Runtime error
Runtime error
update for gpt-j-6b
Browse files
app.py
CHANGED
@@ -49,6 +49,16 @@ conv_template_bf16 = Conversation(
|
|
49 |
sep2="</s>",
|
50 |
)
|
51 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
|
53 |
def set_global_vars(controller_url_, enable_moderation_):
|
54 |
global controller_url, enable_moderation
|
@@ -236,7 +246,7 @@ def http_bot(state, model_selector, temperature, max_new_tokens, request: gr.Req
|
|
236 |
"prompt": prompt,
|
237 |
"temperature": temperature,
|
238 |
"max_new_tokens": max_new_tokens,
|
239 |
-
"stop": "
|
240 |
}
|
241 |
logger.info(f"==== request ====\n{pload}")
|
242 |
|
|
|
49 |
sep2="</s>",
|
50 |
)
|
51 |
|
52 |
+
conv_template_bf16 = Conversation(
|
53 |
+
system="",
|
54 |
+
roles=("", ""),
|
55 |
+
messages=(),
|
56 |
+
offset=0,
|
57 |
+
sep_style=SeparatorStyle.OASST_PYTHIA,
|
58 |
+
sep=" ",
|
59 |
+
sep2="<|endoftext|>",
|
60 |
+
)
|
61 |
+
|
62 |
|
63 |
def set_global_vars(controller_url_, enable_moderation_):
|
64 |
global controller_url, enable_moderation
|
|
|
246 |
"prompt": prompt,
|
247 |
"temperature": temperature,
|
248 |
"max_new_tokens": max_new_tokens,
|
249 |
+
"stop": "<|endoftext|>"
|
250 |
}
|
251 |
logger.info(f"==== request ====\n{pload}")
|
252 |
|