NegBioDB / tests /test_ct_llm_eval.py
jang1563's picture
NegBioDB final: 4 domains, fully audited
6d1bbc7
"""Tests for CT LLM evaluation functions (src/negbiodb_ct/llm_eval.py)."""
import json
import pytest
from negbiodb_ct.llm_eval import (
CT_EVIDENCE_KEYWORDS,
CT_L2_REQUIRED_FIELDS,
CT_L3_JUDGE_PROMPT,
compute_all_ct_llm_metrics,
evaluate_ct_l1,
evaluate_ct_l2,
evaluate_ct_l3,
evaluate_ct_l4,
parse_ct_l1_answer,
parse_ct_l2_response,
parse_ct_l3_judge_scores,
parse_ct_l4_answer,
)
# ── CT-L1 Parser Tests ───────────────────────────────────────────────────
class TestParseCTL1Answer:
def test_single_letter_a_through_e(self):
for letter in "ABCDE":
assert parse_ct_l1_answer(letter) == letter
def test_case_insensitive(self):
assert parse_ct_l1_answer("a") == "A"
assert parse_ct_l1_answer("e") == "E"
def test_answer_colon_format(self):
assert parse_ct_l1_answer("Answer: E") == "E"
assert parse_ct_l1_answer("Answer: B") == "B"
def test_answer_is_format(self):
assert parse_ct_l1_answer("The answer is C") == "C"
def test_parenthesized(self):
assert parse_ct_l1_answer("(D) Strategic discontinuation") == "D"
def test_letter_dot_format(self):
assert parse_ct_l1_answer("A. Safety issue") == "A"
def test_letter_with_explanation(self):
assert parse_ct_l1_answer("B\nThis trial failed to show efficacy") == "B"
def test_empty_returns_none(self):
assert parse_ct_l1_answer("") is None
def test_no_valid_letter_returns_none(self):
assert parse_ct_l1_answer("I don't know the answer") is None
def test_e_category_recognized(self):
"""CT uses E (5-way) unlike DTI (4-way A-D)."""
assert parse_ct_l1_answer("E") == "E"
assert parse_ct_l1_answer("Answer: E") == "E"
assert parse_ct_l1_answer("(E)") == "E"
# ── CT-L1 Evaluator Tests ───────────────────────────────────────────────
class TestEvaluateCTL1:
def test_perfect_accuracy(self):
preds = ["A", "B", "C", "D", "E"]
golds = ["A", "B", "C", "D", "E"]
result = evaluate_ct_l1(preds, golds)
assert result["accuracy"] == 1.0
assert result["parse_rate"] == 1.0
def test_zero_accuracy(self):
preds = ["B", "A", "D", "C", "A"]
golds = ["A", "B", "C", "D", "E"]
result = evaluate_ct_l1(preds, golds)
assert result["accuracy"] == 0.0
def test_per_class_accuracy(self):
preds = ["A", "B", "C"]
golds = ["A", "B", "D"]
classes = ["safety", "efficacy", "strategic"]
result = evaluate_ct_l1(preds, golds, gold_classes=classes)
assert "per_class_accuracy" in result
assert result["per_class_accuracy"]["safety"] == 1.0
assert result["per_class_accuracy"]["strategic"] == 0.0
def test_per_difficulty_accuracy(self):
preds = ["A", "A", "B", "B"]
golds = ["A", "B", "B", "A"]
diffs = ["easy", "easy", "hard", "hard"]
result = evaluate_ct_l1(preds, golds, difficulties=diffs)
assert "per_difficulty_accuracy" in result
assert result["per_difficulty_accuracy"]["easy"] == 0.5
assert result["per_difficulty_accuracy"]["hard"] == 0.5
def test_parse_failures(self):
preds = ["A", "no valid response here at all", "C"]
golds = ["A", "B", "C"]
result = evaluate_ct_l1(preds, golds)
assert result["n_valid"] == 2
assert result["parse_rate"] == pytest.approx(2 / 3)
assert result["accuracy"] == 1.0 # Both parsed correctly
def test_empty_predictions(self):
result = evaluate_ct_l1([], [])
assert result["accuracy"] == 0.0
assert result["n_total"] == 0
def test_all_unparseable(self):
preds = ["xyz", "hello", "???"]
golds = ["A", "B", "C"]
result = evaluate_ct_l1(preds, golds)
assert result["accuracy"] == 0.0
assert result["n_valid"] == 0
# ── CT-L2 Parser Tests ──────────────────────────────────────────────────
class TestParseCTL2Response:
def test_valid_json(self):
obj = {"failure_category": "efficacy", "failure_subcategory": "futility"}
result = parse_ct_l2_response(json.dumps(obj))
assert result == obj
def test_json_with_code_fences(self):
raw = '```json\n{"failure_category": "safety"}\n```'
result = parse_ct_l2_response(raw)
assert result["failure_category"] == "safety"
def test_json_embedded_in_text(self):
raw = 'Here is the result: {"failure_category": "enrollment"} as expected.'
result = parse_ct_l2_response(raw)
assert result["failure_category"] == "enrollment"
def test_invalid_json(self):
assert parse_ct_l2_response("not json at all") is None
def test_partial_fields(self):
obj = {"failure_category": "safety"} # Missing other fields
result = parse_ct_l2_response(json.dumps(obj))
assert result is not None
assert result["failure_category"] == "safety"
# ── CT-L2 Evaluator Tests ───────────────────────────────────────────────
class TestEvaluateCTL2:
def test_perfect_schema_compliance(self):
pred_obj = {f: "test" for f in CT_L2_REQUIRED_FIELDS}
pred_obj["quantitative_evidence"] = True
preds = [json.dumps(pred_obj)]
golds = [{"gold_answer": "efficacy", "failure_category": "efficacy"}]
result = evaluate_ct_l2(preds, golds)
assert result["schema_compliance"] == 1.0
assert result["parse_rate"] == 1.0
def test_category_accuracy(self):
pred_obj = {"failure_category": "safety", "failure_subcategory": "toxicity",
"affected_system": "liver", "severity_indicator": "severe",
"quantitative_evidence": False, "decision_maker": "dsmb",
"patient_impact": "hepatic injury"}
preds = [json.dumps(pred_obj)]
golds = [{"gold_answer": "safety"}]
result = evaluate_ct_l2(preds, golds)
assert result["category_accuracy"] == 1.0
def test_wrong_category(self):
pred_obj = {"failure_category": "efficacy"}
preds = [json.dumps(pred_obj)]
golds = [{"gold_answer": "safety"}]
result = evaluate_ct_l2(preds, golds)
assert result["category_accuracy"] == 0.0
def test_parse_rate(self):
preds = ['{"failure_category": "safety"}', "not json", '{"failure_category": "efficacy"}']
golds = [{"gold_answer": "safety"}, {"gold_answer": "efficacy"}, {"gold_answer": "efficacy"}]
result = evaluate_ct_l2(preds, golds)
assert result["parse_rate"] == pytest.approx(2 / 3)
def test_empty_predictions(self):
result = evaluate_ct_l2([], [])
assert result["n_total"] == 0
# ── CT-L3 Judge Score Parser Tests ───────────────────────────────────────
class TestParseCTL3JudgeScores:
def test_valid_scores(self):
resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5, "specificity": 2})
scores = parse_ct_l3_judge_scores(resp)
assert scores == {"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0}
def test_out_of_range(self):
resp = json.dumps({"accuracy": 6, "reasoning": 0, "completeness": 3, "specificity": 3})
scores = parse_ct_l3_judge_scores(resp)
assert scores is None # 6 and 0 are out of range
def test_missing_dimension(self):
resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5})
scores = parse_ct_l3_judge_scores(resp)
assert scores is None # specificity missing
def test_invalid_json(self):
scores = parse_ct_l3_judge_scores("not json")
assert scores is None
# ── CT-L3 Evaluator Tests ───────────────────────────────────────────────
class TestEvaluateCTL3:
def test_aggregation(self):
scores = [
{"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0},
{"accuracy": 2.0, "reasoning": 5.0, "completeness": 3.0, "specificity": 4.0},
]
result = evaluate_ct_l3(scores)
assert result["accuracy"]["mean"] == pytest.approx(3.0)
assert result["reasoning"]["mean"] == pytest.approx(4.0)
assert result["overall"]["mean"] == pytest.approx(3.5)
assert result["n_valid"] == 2
def test_none_handling(self):
scores = [
{"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0},
None,
]
result = evaluate_ct_l3(scores)
assert result["n_valid"] == 1
assert result["n_total"] == 2
def test_all_none(self):
result = evaluate_ct_l3([None, None])
assert result["n_valid"] == 0
assert result["accuracy"]["mean"] == 0.0
def test_empty(self):
result = evaluate_ct_l3([])
assert result["n_valid"] == 0
# ── CT-L4 Parser Tests ──────────────────────────────────────────────────
class TestParseCTL4Answer:
def test_tested(self):
answer, evidence = parse_ct_l4_answer("tested\nNCT01234567 completed in 2020")
assert answer == "tested"
assert "NCT01234567" in evidence
def test_untested(self):
answer, evidence = parse_ct_l4_answer("untested\nNo registered trials found")
assert answer == "untested"
assert "No registered" in evidence
def test_not_tested_variant(self):
answer, _ = parse_ct_l4_answer("not tested\nReasoning...")
assert answer == "untested"
def test_not_been_tested_variant(self):
answer, _ = parse_ct_l4_answer("This combination has not been tested\nEvidence...")
assert answer == "untested"
def test_never_been_tested_variant(self):
answer, _ = parse_ct_l4_answer("This drug-disease combination has never been tested in a trial.")
assert answer == "untested"
def test_no_evidence(self):
answer, evidence = parse_ct_l4_answer("tested")
assert answer == "tested"
assert evidence is None
def test_empty(self):
answer, evidence = parse_ct_l4_answer("")
assert answer is None
assert evidence is None
# ── CT-L4 Evaluator Tests ───────────────────────────────────────────────
class TestEvaluateCTL4:
def test_perfect_accuracy(self):
preds = ["tested\nNCT123", "untested\nNo trials"]
golds = ["tested", "untested"]
result = evaluate_ct_l4(preds, golds)
assert result["accuracy"] == 1.0
def test_temporal_pre_2020_post_2023(self):
"""CT uses pre_2020/post_2023, NOT DTI's pre_2023/post_2024."""
preds = ["tested\nNCT001", "tested\nNCT002", "untested\nNone", "tested\nNCT003"]
golds = ["tested", "tested", "untested", "untested"]
temporal = ["pre_2020", "post_2023", "pre_2020", "post_2023"]
result = evaluate_ct_l4(preds, golds, temporal_groups=temporal)
# pre_2020: tested→tested (correct), untested→untested (correct) → 100%
assert result["accuracy_pre_2020"] == 1.0
# post_2023: tested→tested (correct), untested→tested (wrong) → 50%
assert result["accuracy_post_2023"] == 0.5
def test_contamination_flag(self):
"""Flag when pre_2020 accuracy exceeds post_2023 by >15%."""
preds = ["tested\nA", "tested\nB", "tested\nC", "untested\nD",
"tested\nE", "tested\nF", "tested\nG", "tested\nH"]
golds = ["tested", "tested", "tested", "untested",
"untested", "untested", "untested", "untested"]
temporal = ["pre_2020", "pre_2020", "pre_2020", "pre_2020",
"post_2023", "post_2023", "post_2023", "post_2023"]
result = evaluate_ct_l4(preds, golds, temporal)
# pre_2020: 3 correct + 1 correct = 4/4 = 100%
# post_2023: 0/4 = 0%
assert result["contamination_flag"] is True
assert result["contamination_gap"] == pytest.approx(1.0)
def test_no_contamination(self):
preds = ["tested\nA", "untested\nB"]
golds = ["tested", "untested"]
temporal = ["pre_2020", "post_2023"]
result = evaluate_ct_l4(preds, golds, temporal)
assert result["contamination_flag"] is False
def test_evidence_citation_rate(self):
"""Evidence needs BOTH >50 chars AND domain keyword (AND logic)."""
preds = [
# >50 chars AND contains "nct" keyword → pass
"tested\nTrial NCT01234567 demonstrated that the drug was effective in reducing primary endpoint with p-value 0.003",
# >50 chars but NO keyword → fail
"tested\nI think the drug was probably tested somewhere in a large randomized clinical study recently",
# <50 chars but has keyword → fail
"tested\nNCT01234567 showed results",
]
golds = ["tested", "tested", "tested"]
result = evaluate_ct_l4(preds, golds)
# Only 1 of 3 passes both conditions
assert result["evidence_citation_rate"] == pytest.approx(1 / 3)
def test_ct_evidence_keywords(self):
"""CT-specific keywords differ from DTI."""
assert "nct" in CT_EVIDENCE_KEYWORDS
assert "clinicaltrials" in CT_EVIDENCE_KEYWORDS
assert "fda" in CT_EVIDENCE_KEYWORDS
assert "eudract" in CT_EVIDENCE_KEYWORDS
# DTI keywords should NOT be here
assert "chembl" not in CT_EVIDENCE_KEYWORDS
assert "pubchem" not in CT_EVIDENCE_KEYWORDS
def test_empty(self):
result = evaluate_ct_l4([], [])
assert result["accuracy"] == 0.0
# ── Dispatch Tests ───────────────────────────────────────────────────────
class TestDispatch:
def test_ct_l1_dispatch(self):
preds = ["A", "B"]
gold = [{"gold_answer": "A", "gold_category": "safety", "difficulty": "easy"},
{"gold_answer": "B", "gold_category": "efficacy", "difficulty": "hard"}]
result = compute_all_ct_llm_metrics("ct-l1", preds, gold)
assert result["accuracy"] == 1.0
def test_ct_l4_dispatch(self):
preds = ["tested\nNCT123", "untested\nNone"]
gold = [{"gold_answer": "tested", "temporal_group": "pre_2020"},
{"gold_answer": "untested", "temporal_group": "post_2023"}]
result = compute_all_ct_llm_metrics("ct-l4", preds, gold)
assert result["accuracy"] == 1.0
def test_invalid_task_raises(self):
with pytest.raises(ValueError, match="Unknown task"):
compute_all_ct_llm_metrics("l1", ["A"], [{"gold_answer": "A"}])
def test_ct_l2_dispatch(self):
pred_obj = {"failure_category": "efficacy"}
preds = [json.dumps(pred_obj)]
gold = [{"gold_answer": "efficacy"}]
result = compute_all_ct_llm_metrics("ct-l2", preds, gold)
assert result["category_accuracy"] == 1.0
def test_ct_l3_dispatch(self):
resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5, "specificity": 2})
result = compute_all_ct_llm_metrics("ct-l3", [resp], [{}])
assert result["n_valid"] == 1