Commit ·
439e257
1
Parent(s): 1b0e1c0
Add --messages-until-role to truncate messages for regeneration
Browse files- generate-responses.py +16 -0
generate-responses.py
CHANGED
|
@@ -184,6 +184,7 @@ def main(
|
|
| 184 |
attn_implementation: str = "paged|sdpa",
|
| 185 |
subset: Optional[str] = None,
|
| 186 |
split: str = "train",
|
|
|
|
| 187 |
skip_long_prompts: bool = True,
|
| 188 |
max_samples: Optional[int] = None,
|
| 189 |
hf_token: Optional[str] = None,
|
|
@@ -289,6 +290,14 @@ def main(
|
|
| 289 |
for i, example in enumerate(dataset):
|
| 290 |
if use_messages:
|
| 291 |
messages = example[messages_column]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 292 |
prompt = tokenizer.apply_chat_template(
|
| 293 |
messages, tokenize=False, add_generation_prompt=True
|
| 294 |
)
|
|
@@ -451,6 +460,12 @@ Examples:
|
|
| 451 |
type=str,
|
| 452 |
help="Column containing plain text prompts (alternative to --messages-column)",
|
| 453 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 454 |
parser.add_argument(
|
| 455 |
"--output-column",
|
| 456 |
type=str,
|
|
@@ -539,6 +554,7 @@ Examples:
|
|
| 539 |
model_id=args.model_id,
|
| 540 |
messages_column=args.messages_column,
|
| 541 |
prompt_column=args.prompt_column,
|
|
|
|
| 542 |
output_column=args.output_column,
|
| 543 |
temperature=args.temperature,
|
| 544 |
top_p=args.top_p,
|
|
|
|
| 184 |
attn_implementation: str = "paged|sdpa",
|
| 185 |
subset: Optional[str] = None,
|
| 186 |
split: str = "train",
|
| 187 |
+
messages_until_role: Optional[str] = None,
|
| 188 |
skip_long_prompts: bool = True,
|
| 189 |
max_samples: Optional[int] = None,
|
| 190 |
hf_token: Optional[str] = None,
|
|
|
|
| 290 |
for i, example in enumerate(dataset):
|
| 291 |
if use_messages:
|
| 292 |
messages = example[messages_column]
|
| 293 |
+
# Optionally truncate messages up to last occurrence of a given role
|
| 294 |
+
if messages_until_role:
|
| 295 |
+
truncated = []
|
| 296 |
+
for msg in messages:
|
| 297 |
+
truncated.append(msg)
|
| 298 |
+
if msg["role"] == messages_until_role:
|
| 299 |
+
last_truncated = list(truncated)
|
| 300 |
+
messages = last_truncated
|
| 301 |
prompt = tokenizer.apply_chat_template(
|
| 302 |
messages, tokenize=False, add_generation_prompt=True
|
| 303 |
)
|
|
|
|
| 460 |
type=str,
|
| 461 |
help="Column containing plain text prompts (alternative to --messages-column)",
|
| 462 |
)
|
| 463 |
+
parser.add_argument(
|
| 464 |
+
"--messages-until-role",
|
| 465 |
+
type=str,
|
| 466 |
+
help="Truncate messages up to the last message with this role. "
|
| 467 |
+
"Use 'user' to strip existing assistant responses and regenerate them.",
|
| 468 |
+
)
|
| 469 |
parser.add_argument(
|
| 470 |
"--output-column",
|
| 471 |
type=str,
|
|
|
|
| 554 |
model_id=args.model_id,
|
| 555 |
messages_column=args.messages_column,
|
| 556 |
prompt_column=args.prompt_column,
|
| 557 |
+
messages_until_role=args.messages_until_role,
|
| 558 |
output_column=args.output_column,
|
| 559 |
temperature=args.temperature,
|
| 560 |
top_p=args.top_p,
|