Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| #!/usr/bin/env python3 | |
| """Pre-annotate L2 candidates with LLM extraction for human correction. | |
| Reads l2_candidates.jsonl, runs each abstract through an LLM using the | |
| L2 zero-shot prompt, and saves the pre-annotated results for human review. | |
| Usage: | |
| python scripts/preannotate_l2.py --provider openai --model gpt-4o-mini | |
| python scripts/preannotate_l2.py --provider anthropic --model claude-sonnet-4-6 | |
| Output: | |
| exports/llm_benchmarks/l2_preannotated.jsonl | |
| """ | |
| import argparse | |
| import json | |
| import time | |
| from pathlib import Path | |
| from negbiodb.llm_client import LLMClient | |
| from negbiodb.llm_eval import parse_l2_response | |
| from negbiodb.llm_prompts import SYSTEM_PROMPT, format_l2_prompt | |
| PROJECT_ROOT = Path(__file__).resolve().parent.parent | |
| DATA_DIR = PROJECT_ROOT / "exports" / "llm_benchmarks" | |
| CANDIDATES_FILE = DATA_DIR / "l2_candidates.jsonl" | |
| OUTPUT_FILE = DATA_DIR / "l2_preannotated.jsonl" | |
| def main(): | |
| parser = argparse.ArgumentParser(description="Pre-annotate L2 candidates") | |
| parser.add_argument("--provider", default="openai", | |
| choices=["openai", "gemini", "vllm", "anthropic"]) | |
| parser.add_argument("--model", default="gpt-4o-mini") | |
| parser.add_argument("--api-base", default=None) | |
| parser.add_argument("--api-key", default=None) | |
| parser.add_argument("--temperature", type=float, default=0.0) | |
| parser.add_argument("--max-tokens", type=int, default=2048) | |
| args = parser.parse_args() | |
| # Load candidates | |
| candidates = [] | |
| with open(CANDIDATES_FILE) as f: | |
| for line in f: | |
| candidates.append(json.loads(line)) | |
| print(f"Loaded {len(candidates)} candidates") | |
| # Resume: check existing output | |
| completed = set() | |
| if OUTPUT_FILE.exists(): | |
| with open(OUTPUT_FILE) as f: | |
| for line in f: | |
| rec = json.loads(line) | |
| completed.add(rec["candidate_id"]) | |
| print(f" Resume: {len(completed)} already processed") | |
| remaining = [c for c in candidates if c["candidate_id"] not in completed] | |
| print(f" Remaining: {len(remaining)}") | |
| if not remaining: | |
| print("All candidates already processed.") | |
| return | |
| # Initialize client | |
| print(f"\nInitializing: {args.model} ({args.provider})") | |
| client = LLMClient( | |
| provider=args.provider, | |
| model=args.model, | |
| api_base=args.api_base, | |
| api_key=args.api_key, | |
| temperature=args.temperature, | |
| max_tokens=args.max_tokens, | |
| ) | |
| # Process each candidate | |
| start_time = time.time() | |
| with open(OUTPUT_FILE, "a") as out_f: | |
| for i, candidate in enumerate(remaining): | |
| system, user = format_l2_prompt( | |
| {"abstract_text": candidate["abstract_text"]}, | |
| config="zero-shot", | |
| ) | |
| try: | |
| response = client.generate(user, system) | |
| except Exception as e: | |
| print(f" Error on {candidate['candidate_id']}: {e}") | |
| response = f"ERROR: {e}" | |
| # Parse LLM extraction | |
| extraction = parse_l2_response(response) | |
| output_rec = { | |
| **candidate, | |
| "llm_response": response, | |
| "llm_extraction": extraction, | |
| "preannotation_model": args.model, | |
| } | |
| out_f.write(json.dumps(output_rec, ensure_ascii=False) + "\n") | |
| out_f.flush() | |
| if (i + 1) % 10 == 0: | |
| elapsed = time.time() - start_time | |
| rate = (i + 1) / elapsed * 60 | |
| print(f" Progress: {i + 1}/{len(remaining)} ({rate:.1f}/min)") | |
| elapsed = time.time() - start_time | |
| print(f"\nDone: {len(remaining)} processed in {elapsed:.0f}s") | |
| print(f"Output: {OUTPUT_FILE}") | |
| if __name__ == "__main__": | |
| main() | |