smartgridbench-review-artifact / code /tests /test_orchestration_utils.py
garn-garn's picture
Add anonymized executable code package
881f9f2 verified
from __future__ import annotations
import asyncio
import os
import sys
import tempfile
import unittest
from pathlib import Path
from types import SimpleNamespace
from unittest.mock import patch
REPO_ROOT = Path(__file__).resolve().parents[1]
sys.path.insert(0, str(REPO_ROOT / "scripts"))
from orchestration_utils import ( # noqa: E402
available_sensor_ids,
bootstrap_aob,
build_fault_risk_adjudication_state,
build_executor,
build_parser,
build_planner_descriptions,
build_planning_question,
build_tool_catalog_for_executor,
can_retry_missing_evidence,
canonicalize_step_result,
close_executor,
compact_step_for_context,
compact_verifier_result,
SelfAskDecision,
VerificationDecision,
build_missing_evidence_repair_state,
build_llm,
build_retry_question,
build_suffix_replan_question,
current_missing_evidence_hit,
effective_server_paths,
fault_risk_adjudication_failed_step,
finalize_missing_evidence_repair_state,
generate_suffix_plan,
load_fault_risk_adjudication_config,
load_missing_evidence_repair_config,
load_plan_execute_planner,
maybe_self_ask,
normalize_plan_steps,
normalize_response_text,
parse_json_object,
preflight_aob_runtime_dependencies,
repair_sensor_task_text,
repair_target_key,
response_error_payload,
record_missing_evidence_retry_attempt,
serialize_step_result,
should_skip_invalid_sensor_step,
summarize_answer,
summarize_terminal_failures,
tool_schema_for_step,
verify_step,
)
from mitigation_guards import ( # noqa: E402
EXPLICIT_FAULT_RISK_ADJUDICATION_NAME,
MISSING_EVIDENCE_REPAIR_NAME,
_fault_risk_adjudication_applies,
apply_explicit_fault_risk_adjudication,
apply_missing_evidence_final_answer_guard,
build_explicit_fault_risk_adjudication,
env_flag_enabled,
scan_missing_evidence,
)
class DummyLLM:
def __init__(self, responses):
self._responses = list(responses)
self.prompts = []
def generate(self, prompt: str, **_kwargs) -> str:
self.prompts.append(prompt)
if not self._responses:
raise AssertionError(f"No response left for prompt: {prompt[:120]}")
return self._responses.pop(0)
class OrchestrationUtilsTests(unittest.TestCase):
def test_build_planning_question_appends_guardrails(self):
question = build_planning_question("Investigate transformer T-015.")
self.assertIn("Use only these real servers", question)
self.assertIn("get_dga_record and analyze_dga are FMSR tools", question)
def test_build_planner_descriptions_adds_repo_specific_notes(self):
descriptions = {
"fmsr": " - get_dga_record(transformer_id: string): Retrieve DGA"
}
tool_catalog = {
"fmsr": {
"get_dga_record": {
"description": "Retrieve DGA",
"schema": "transformer_id: string (transformer id like T-015)",
}
}
}
rendered = build_planner_descriptions(descriptions, tool_catalog)["fmsr"]
self.assertIn("Planner notes:", rendered)
self.assertIn("get_sensor_correlation only", rendered)
self.assertIn("Canonical tool signatures:", rendered)
def test_build_parser_reads_mcp_mode_environment_default(self):
with patch.dict(os.environ, {"MCP_MODE": "optimized"}):
parser = build_parser("runner", "Test runner.")
args = parser.parse_args(["Investigate transformer T-015."])
self.assertEqual(args.mcp_mode, "optimized")
def test_build_llm_honors_max_tokens_without_aob_usage_class(self):
captured = {}
def completion(**kwargs):
captured.update(kwargs)
return SimpleNamespace(
choices=[SimpleNamespace(message=SimpleNamespace(content="ok"))],
usage=SimpleNamespace(prompt_tokens=3, completion_tokens=2),
)
fake_litellm = SimpleNamespace(completion=completion)
with patch.dict(
os.environ,
{
"MAX_TOKENS": "17",
"LITELLM_API_KEY": "key",
"LITELLM_BASE_URL": "http://127.0.0.1:8000",
},
), patch.dict(sys.modules, {"litellm": fake_litellm}):
llm = build_llm("openai/test-model")
result = llm.generate_with_usage("prompt")
self.assertEqual(result.text, "ok")
self.assertEqual(result.input_tokens, 3)
self.assertEqual(result.output_tokens, 2)
self.assertEqual(captured["max_tokens"], 17)
def test_build_tool_catalog_for_executor_prefers_executor_cache(self):
class Executor:
def __init__(self):
self.called = False
async def get_tool_catalog(self):
self.called = True
return {"iot": {"list_assets": {"description": "", "schema": ""}}}
executor = Executor()
result = asyncio.run(
build_tool_catalog_for_executor(
executor,
{"iot": REPO_ROOT / "does_not_need_to_exist.py"},
)
)
self.assertTrue(executor.called)
self.assertIn("list_assets", result["iot"])
def test_build_executor_skips_aob_agent_package_side_effects(self):
with tempfile.TemporaryDirectory() as tmp:
aob_path = Path(tmp)
agent_path = aob_path / "src" / "agent"
plan_execute_path = agent_path / "plan_execute"
plan_execute_path.mkdir(parents=True)
(agent_path / "__init__.py").write_text(
'raise RuntimeError("agent package init should not run")\n',
encoding="utf-8",
)
(agent_path / "runner.py").write_text(
"DEFAULT_SERVER_PATHS = {'iot': 'iot.py'}\n",
encoding="utf-8",
)
(plan_execute_path / "__init__.py").write_text("", encoding="utf-8")
(plan_execute_path / "executor.py").write_text(
"from ..runner import DEFAULT_SERVER_PATHS\n\n"
"class Executor:\n"
" def __init__(self, llm, server_paths):\n"
" self.llm = llm\n"
" self.server_paths = server_paths\n"
" self.default_server_paths = DEFAULT_SERVER_PATHS\n",
encoding="utf-8",
)
old_agent_modules = {
name: module
for name, module in sys.modules.items()
if name == "agent" or name.startswith("agent.")
}
old_sys_path = list(sys.path)
for name in old_agent_modules:
sys.modules.pop(name, None)
try:
bootstrap_aob(aob_path)
executor = build_executor(
object(),
{"iot": Path("iot.py")},
mcp_mode="baseline",
)
finally:
for name in list(sys.modules):
if name == "agent" or name.startswith("agent."):
sys.modules.pop(name, None)
sys.modules.update(old_agent_modules)
sys.path[:] = old_sys_path
self.assertEqual(executor.server_paths, {"iot": Path("iot.py")})
self.assertEqual(executor.default_server_paths, {"iot": "iot.py"})
def test_load_plan_execute_planner_accepts_task_dependencies(self):
with tempfile.TemporaryDirectory() as tmp:
aob_path = Path(tmp)
agent_path = aob_path / "src" / "agent"
plan_execute_path = agent_path / "plan_execute"
plan_execute_path.mkdir(parents=True)
(agent_path / "__init__.py").write_text(
'raise RuntimeError("agent package init should not run")\n',
encoding="utf-8",
)
(plan_execute_path / "__init__.py").write_text("", encoding="utf-8")
(plan_execute_path / "planner.py").write_text(
"import re\n"
"_DEP_NUM_RE = re.compile(r'#S(\\d+)')\n\n"
"class Planner:\n"
" pass\n",
encoding="utf-8",
)
old_agent_modules = {
name: module
for name, module in sys.modules.items()
if name == "agent" or name.startswith("agent.")
}
old_sys_path = list(sys.path)
for name in old_agent_modules:
sys.modules.pop(name, None)
try:
bootstrap_aob(aob_path)
Planner = load_plan_execute_planner()
from agent.plan_execute import planner as planner_module
finally:
for name in list(sys.modules):
if name == "agent" or name.startswith("agent."):
sys.modules.pop(name, None)
sys.modules.update(old_agent_modules)
sys.path[:] = old_sys_path
self.assertEqual(Planner.__name__, "Planner")
self.assertEqual(planner_module._DEP_NUM_RE.findall("#Task1 #S2"), ["1", "2"])
def test_close_executor_awaits_optional_aclose(self):
class Executor:
def __init__(self):
self.closed = False
async def aclose(self):
self.closed = True
executor = Executor()
asyncio.run(close_executor(executor))
self.assertTrue(executor.closed)
def test_normalize_plan_steps_reroutes_unique_tool_server(self):
class Step:
def __init__(self, step_number, task, server, tool, dependencies=None):
self.step_number = step_number
self.task = task
self.server = server
self.tool = tool
self.tool_args = {}
self.dependencies = dependencies or []
self.expected_output = ""
class Plan:
def __init__(self, steps):
self.steps = steps
plan = Plan(
[
Step(1, "Fetch the latest DGA record", "iot", "get_dga_record"),
Step(2, "Analyze the gases", "iot", "analyze_dga", [1]),
]
)
tool_catalog = {
"iot": {"get_asset_metadata": {"description": "", "schema": ""}},
"fmsr": {
"get_dga_record": {"description": "", "schema": ""},
"analyze_dga": {"description": "", "schema": ""},
},
}
warnings = normalize_plan_steps(plan, tool_catalog)
self.assertEqual(plan.steps[0].server, "fmsr")
self.assertEqual(plan.steps[1].server, "fmsr")
self.assertEqual(plan.steps[1].dependencies, [1])
self.assertTrue(any("Rerouted step 1" in warning for warning in warnings))
def test_normalize_plan_steps_accepts_short_server_aliases(self):
class Step:
def __init__(self, step_number, task, server, tool, dependencies=None):
self.step_number = step_number
self.task = task
self.server = server
self.tool = tool
self.tool_args = {}
self.dependencies = dependencies or []
self.expected_output = ""
class Plan:
def __init__(self, steps):
self.steps = steps
plan = Plan([Step(1, "List transformer sensors", "i", "list_sensors")])
tool_catalog = {
"iot": {"list_sensors": {"description": "", "schema": ""}},
}
warnings = normalize_plan_steps(plan, tool_catalog)
self.assertEqual(plan.steps[0].server, "iot")
self.assertTrue(
any("Normalized server alias" in warning for warning in warnings)
)
def test_normalize_plan_steps_drops_terminal_none_step(self):
class Step:
def __init__(self, step_number, task, server, tool, dependencies=None):
self.step_number = step_number
self.task = task
self.server = server
self.tool = tool
self.tool_args = {}
self.dependencies = dependencies or []
self.expected_output = ""
class Plan:
def __init__(self, steps):
self.steps = steps
plan = Plan(
[
Step(1, "Fetch DGA record", "fmsr", "get_dga_record"),
Step(2, "Summarize findings", "none", "none", [1]),
]
)
tool_catalog = {"fmsr": {"get_dga_record": {"description": "", "schema": ""}}}
warnings = normalize_plan_steps(plan, tool_catalog)
self.assertEqual(len(plan.steps), 1)
self.assertEqual(plan.steps[0].step_number, 1)
self.assertTrue(
any(
"Dropped terminal synthesis-only step" in warning
for warning in warnings
)
)
def test_tool_schema_for_step_uses_hintful_schema(self):
tool_catalog = {
"fmsr": {
"get_sensor_correlation": {
"description": "",
"schema": "failure_mode_id: string (failure mode id like FM-006; not a transformer id)",
}
}
}
schema = tool_schema_for_step(tool_catalog, "fmsr", "get_sensor_correlation")
self.assertIn("FM-006", schema)
self.assertIn("not a transformer id", schema)
def test_normalize_response_text_wraps_newline_json_objects_into_list(self):
response = '{"sensor_id":"load_current_a"}\n{"sensor_id":"winding_temp_top_c"}'
normalized = normalize_response_text(response)
self.assertIn("[", normalized)
self.assertIn("winding_temp_top_c", normalized)
def test_available_sensor_ids_reads_list_sensors_context(self):
class Result:
tool = "list_sensors"
response = (
'[{"sensor_id":"load_current_a"},{"sensor_id":"winding_temp_top_c"}]'
)
sensors = available_sensor_ids({3: Result()})
self.assertEqual(sensors, {"load_current_a", "winding_temp_top_c"})
def test_repair_sensor_task_text_rewrites_known_alias(self):
task, warning = repair_sensor_task_text(
"Compute the trend of the winding_temp_c sensor readings.",
{"winding_temp_top_c", "oil_temp_c"},
)
self.assertIn("winding_temp_top_c", task)
self.assertIn("winding_temp_c -> winding_temp_top_c", warning)
def test_should_skip_invalid_sensor_step_for_iot_dga_lookup(self):
class Step:
tool = "get_sensor_readings"
task = "Get time-series readings for sensor dga_h2_ppm on transformer T-015"
should_skip, reason = should_skip_invalid_sensor_step(
Step(), {"load_current_a", "winding_temp_top_c"}
)
self.assertTrue(should_skip)
self.assertIn("DGA data in this repo comes from FMSR", reason)
def test_canonicalize_step_result_normalizes_response(self):
class Result:
response = (
'{"sensor_id":"load_current_a"}\n{"sensor_id":"winding_temp_top_c"}'
)
result = Result()
canonicalize_step_result(result)
self.assertIn("[", result.response)
self.assertIn("winding_temp_top_c", result.response)
def test_compact_step_for_context_truncates_large_response_and_error(self):
class Result:
response = "A" * 3000
error = "B" * 1200
result = Result()
compact_step_for_context(result)
self.assertIn("[truncated", result.response)
self.assertIn("[truncated", result.error)
self.assertLess(len(result.response), 1400)
self.assertLess(len(result.error), 800)
def test_compact_verifier_result_truncates_large_response(self):
compacted = compact_verifier_result(
{
"step": 4,
"task": "Get sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"success": True,
"response": "A" * 3000,
"error": None,
}
)
self.assertIn("[truncated", compacted["response"])
self.assertLess(len(compacted["response"]), 1400)
def test_parse_json_object_handles_markdown_fence(self):
raw = '```json\n{"decision": "continue", "reason": "looks good"}\n```'
payload = parse_json_object(raw)
self.assertEqual(payload["decision"], "continue")
def test_parse_json_object_preserves_json_substring_in_payload(self):
raw = '```json\n{"name": "json_value"}\n```'
payload = parse_json_object(raw)
self.assertEqual(payload["name"], "json_value")
def test_response_error_payload_detects_top_level_json_error(self):
self.assertEqual(
response_error_payload('{"error": "Failure mode not found."}'),
"Failure mode not found.",
)
self.assertIsNone(response_error_payload('{"result": "ok"}'))
def test_response_error_payload_detects_list_wrapped_error(self):
self.assertEqual(
response_error_payload([{"error": "No sensor data found for 'T-020'."}]),
"No sensor data found for 'T-020'.",
)
def test_response_error_payload_detects_unknown_tool_string(self):
self.assertEqual(
response_error_payload("Unknown tool: get_dga_record"),
"Unknown tool: get_dga_record",
)
def test_serialize_step_result_marks_json_error_payload_as_failure(self):
class Result:
step_number = 4
task = "Infer probable fault mode"
server = "fmsr"
tool = "get_sensor_correlation"
tool_args = {"failure_mode_id": "T-015"}
response = '{"error": "Failure mode \\"T-015\\" not found."}'
error = None
success = True
payload = serialize_step_result(Result())
self.assertFalse(payload["success"])
self.assertTrue(payload["executor_success"])
self.assertEqual(payload["error"], 'Failure mode "T-015" not found.')
def test_serialize_step_result_marks_unknown_tool_string_as_failure(self):
class Result:
step_number = 1
task = "Get DGA record"
server = "iot"
tool = "get_dga_record"
tool_args = {"transformer_id": "T-020"}
response = "Unknown tool: get_dga_record"
error = None
success = True
payload = serialize_step_result(Result())
self.assertFalse(payload["success"])
self.assertTrue(payload["executor_success"])
self.assertEqual(payload["error"], "Unknown tool: get_dga_record")
def test_serialize_step_result_carries_runner_repair_metadata(self):
class Result:
step_number = 2
task = "Inspect DGA sensor"
server = "iot"
tool = "get_sensor_readings"
tool_args = {"sensor_id": "dga_h2_ppm"}
response = "Skipped invalid IoT DGA lookup."
error = None
success = True
runner_repair = "invalid_iot_dga_sensor_lookup"
runner_repair_reason = "Use FMSR DGA tools instead."
payload = serialize_step_result(Result())
self.assertTrue(payload["success"])
self.assertEqual(payload["runner_repair"], "invalid_iot_dga_sensor_lookup")
self.assertEqual(payload["runner_repair_reason"], "Use FMSR DGA tools instead.")
def test_maybe_self_ask_defaults_to_original_question_when_false(self):
llm = DummyLLM(
[
'{"needs_self_ask": false, "clarifying_questions": [], "assumptions": [], "augmented_question": "ignored"}'
]
)
decision = maybe_self_ask("What failed at site MAIN?", llm)
self.assertEqual(
decision,
SelfAskDecision(
needs_self_ask=False,
clarifying_questions=[],
assumptions=[],
augmented_question="What failed at site MAIN?",
),
)
def test_maybe_self_ask_constructs_augmented_question_when_missing(self):
llm = DummyLLM(
[
'{"needs_self_ask": true, "clarifying_questions": ["Which transformer?"], "assumptions": ["Use the most recent incident"], "augmented_question": ""}'
]
)
decision = maybe_self_ask("Investigate the overheating alert.", llm)
self.assertTrue(decision.needs_self_ask)
self.assertIn("Which transformer?", decision.augmented_question)
self.assertIn("Use the most recent incident", decision.augmented_question)
def test_build_suffix_replan_question_includes_focus_and_remaining_steps(self):
class Step:
def __init__(self, step_number, task, server, tool):
self.step_number = step_number
self.task = task
self.server = server
self.tool = tool
history = [
{
"step": 1,
"task": "Find the transformer",
"server": "iot",
"tool": "get_asset_metadata",
"response": "Transformer is T-015.",
"error": None,
"success": True,
}
]
remaining = [Step(2, "Fetch DGA record", "fmsr", "get_dga_record")]
question = build_suffix_replan_question(
"Investigate T-015.",
"Investigate T-015.",
history,
remaining,
VerificationDecision(
decision="replan_suffix",
reason="The investigation should pivot to thermal fault confirmation.",
updated_focus="Confirm the thermal fault before creating the work order.",
),
)
self.assertIn("thermal fault", question)
self.assertIn("Fetch DGA record", question)
self.assertIn("Dependency1: None", question)
self.assertIn("never on completed steps", question)
def test_build_retry_question_includes_previous_attempt_and_reason(self):
class Step:
def __init__(self, step_number, task):
self.step_number = step_number
self.task = task
question = build_retry_question(
"Investigate T-015.",
"Investigate T-015 with internal clarifications.",
Step(2, "Fetch DGA record"),
{"response": "Returned the wrong asset record."},
VerificationDecision(
decision="retry",
reason="The tool arguments targeted the wrong asset.",
updated_focus="Use the transformer identified in step 1.",
),
retries_used=1,
)
self.assertIn("wrong asset", question)
self.assertIn("Retry attempt number: 1", question)
self.assertIn("step 1", question.lower())
def test_summarize_answer_uses_only_terminal_history_entries(self):
llm = DummyLLM(["Final answer."])
answer = summarize_answer(
"What happened?",
[
{
"step": 1,
"task": "Lookup alert",
"server": "iot",
"response": "Old retry response",
"error": None,
"success": True,
},
{
"step": 1,
"task": "Lookup alert",
"server": "iot",
"response": "Final step 1 response",
"error": None,
"success": True,
},
{
"step": 2,
"task": "Create WO",
"server": "wo",
"response": "WO created",
"error": None,
"success": True,
},
],
llm,
)
self.assertEqual(answer, "Final answer.")
self.assertIn("Final step 1 response", llm.prompts[0])
self.assertNotIn("Old retry response", llm.prompts[0])
def test_summarize_answer_falls_back_when_llm_raises(self):
class BrokenLLM:
def generate(self, _prompt: str, **_kwargs) -> str:
raise RuntimeError("context window exceeded")
with self.assertLogs("orchestration_utils", level="WARNING") as logs:
answer = summarize_answer(
"What happened?",
[
{
"step": 1,
"task": "Get readings",
"server": "iot",
"tool": "get_sensor_readings",
"response": "A" * 2000,
"error": None,
"success": True,
},
{
"step": 2,
"task": "Create WO",
"server": "wo",
"tool": "create_work_order",
"response": None,
"error": "could not create",
"success": False,
},
],
BrokenLLM(),
)
self.assertIn("Compact terminal results", answer)
self.assertIn("create_work_order", answer)
self.assertTrue(any("summarization failed" in line for line in logs.output))
def test_summarize_terminal_failures_only_keeps_terminal_failed_steps(self):
failures = summarize_terminal_failures(
[
{
"step": 1,
"task": "Lookup alert",
"server": "iot",
"tool": "get_alert",
"error": "transient failure",
"success": False,
},
{
"step": 1,
"task": "Lookup alert",
"server": "iot",
"tool": "get_alert",
"error": None,
"success": True,
},
{
"step": 2,
"task": "Create WO",
"server": "wo",
"tool": "create_work_order",
"error": "Unknown server",
"success": False,
"verifier_decision": "error",
},
]
)
self.assertEqual(
failures,
[
{
"step": 2,
"task": "Create WO",
"server": "wo",
"tool": "create_work_order",
"error": "Unknown server",
"verifier_decision": "error",
}
],
)
def test_generate_suffix_plan_returns_warning_payload_on_invalid_plan(self):
class DummyPlanner:
def generate_plan(self, _question, _descriptions):
raise ValueError("Invalid dependency reference for step 1: #S1")
with self.assertLogs("orchestration_utils", level="WARNING") as logs:
plan, error = generate_suffix_plan(
DummyPlanner(),
"Replan only the suffix.",
{"iot": "list_sensors"},
)
self.assertIsNone(plan)
self.assertEqual(error, "Invalid dependency reference for step 1: #S1")
self.assertIn(
"continuing with the original remaining plan", "\n".join(logs.output)
)
def test_verify_step_warns_and_falls_back_when_json_missing(self):
class Step:
def __init__(self):
self.step_number = 2
self.task = "Fetch DGA record"
self.server = "fmsr"
self.tool = "get_dga_record"
llm = DummyLLM(["not json at all"])
with self.assertLogs("orchestration_utils", level="WARNING") as logs:
decision = verify_step(
"Investigate T-015.",
"Investigate T-015.",
Step(),
{"response": "Some result", "success": True},
[],
[],
llm,
)
self.assertEqual(decision.decision, "continue")
self.assertTrue(any("no parseable JSON" in line for line in logs.output))
def test_verify_step_warns_and_falls_back_when_decision_unknown(self):
class Step:
def __init__(self):
self.step_number = 2
self.task = "Fetch DGA record"
self.server = "fmsr"
self.tool = "get_dga_record"
llm = DummyLLM(
[
'{"decision": "repair_everything", "reason": "nonsense", "updated_focus": ""}'
]
)
with self.assertLogs("orchestration_utils", level="WARNING") as logs:
decision = verify_step(
"Investigate T-015.",
"Investigate T-015.",
Step(),
{"response": "Some result", "success": True},
[],
[],
llm,
)
self.assertEqual(decision.decision, "continue")
self.assertTrue(any("unknown decision" in line for line in logs.output))
def test_verify_step_warns_and_falls_back_when_llm_raises(self):
class Step:
def __init__(self):
self.step_number = 4
self.task = "Get sensor readings"
self.server = "iot"
self.tool = "get_sensor_readings"
class BrokenLLM:
def generate(self, _prompt: str, **_kwargs) -> str:
raise RuntimeError("context window exceeded")
with self.assertLogs("orchestration_utils", level="WARNING") as logs:
decision = verify_step(
"Investigate T-015.",
"Investigate T-015.",
Step(),
{"response": "A" * 5000, "success": True},
[],
[],
BrokenLLM(),
)
self.assertEqual(decision.decision, "continue")
self.assertTrue(any("Verifier failed" in line for line in logs.output))
def test_effective_server_paths_exits_cleanly_on_bad_override(self):
with self.assertRaises(SystemExit) as exc:
effective_server_paths(["bad-entry"], REPO_ROOT)
self.assertIn("Invalid --server entry", str(exc.exception))
def test_preflight_aob_runtime_dependencies_reports_missing_modules(self):
def fake_import_module(name: str):
if name == "litellm":
raise ModuleNotFoundError("No module named 'litellm'")
return object()
with patch(
"orchestration_utils.importlib.import_module",
side_effect=fake_import_module,
):
with self.assertRaises(RuntimeError) as exc:
preflight_aob_runtime_dependencies()
self.assertIn("litellm", str(exc.exception))
self.assertIn("requirements-insomnia.txt", str(exc.exception))
def test_env_flag_enabled_accepts_standard_truthy_values(self):
self.assertTrue(env_flag_enabled("1"))
self.assertTrue(env_flag_enabled("true"))
self.assertFalse(env_flag_enabled("0"))
self.assertFalse(env_flag_enabled(""))
def test_missing_evidence_guard_blocks_substantive_final_answer(self):
payload = {
"answer": "Schedule immediate maintenance for transformer T-015.",
"success": True,
"failed_steps": [],
"history": [
{
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {"sensor_id": "winding_temp_top_c"},
"response": '{"readings": []}',
"error": None,
"success": True,
}
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertFalse(guarded["success"])
self.assertIn("Mitigation guard blocked", guarded["answer"])
self.assertTrue(guarded["mitigation_guard"]["triggered"])
self.assertTrue(guarded["mitigation_guard"]["blocked_final_answer"])
self.assertEqual(
guarded["failed_steps"][-1]["tool"],
"missing_evidence_final_answer_guard",
)
def test_scan_missing_evidence_returns_hit_without_mutating_payload(self):
payload = {
"answer": "Schedule immediate maintenance for transformer T-015.",
"success": True,
"history": [
{
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": '{"readings": []}',
"error": None,
"success": True,
}
],
}
scan = scan_missing_evidence(payload)
self.assertTrue(scan["triggered"])
self.assertEqual(scan["hits"][0]["step"], 1)
self.assertEqual(scan["hits"][0]["tool"], "get_sensor_readings")
self.assertEqual(
scan["hits"][0]["target"],
{"transformer_id": "T-015", "sensor_id": "winding_temp_top_c"},
)
self.assertNotIn("mitigation_guard", payload)
def test_repair_config_requires_missing_evidence_guard(self):
with patch.dict(
os.environ,
{
"ENABLE_MISSING_EVIDENCE_REPAIR": "1",
"ENABLE_MISSING_EVIDENCE_GUARD": "0",
},
):
with self.assertRaisesRegex(
RuntimeError,
"requires ENABLE_MISSING_EVIDENCE_GUARD",
):
load_missing_evidence_repair_config()
def test_repair_config_ignores_budget_env_when_disabled(self):
with patch.dict(
os.environ,
{
"ENABLE_MISSING_EVIDENCE_REPAIR": "0",
"ENABLE_MISSING_EVIDENCE_GUARD": "0",
"MISSING_EVIDENCE_REPAIR_MAX_ATTEMPTS": "0",
"MISSING_EVIDENCE_REPAIR_MAX_ATTEMPTS_PER_TARGET": "not-an-int",
},
):
config = load_missing_evidence_repair_config()
self.assertFalse(config.enabled)
self.assertEqual(config.max_attempts, 0)
self.assertEqual(config.max_attempts_per_target, 0)
def test_adjudication_config_requires_missing_evidence_guard(self):
with patch.dict(
os.environ,
{
"ENABLE_EXPLICIT_FAULT_RISK_ADJUDICATION": "1",
"ENABLE_MISSING_EVIDENCE_GUARD": "0",
},
):
with self.assertRaisesRegex(
RuntimeError,
"requires ENABLE_MISSING_EVIDENCE_GUARD",
):
load_fault_risk_adjudication_config()
def test_fault_risk_adjudication_finalizes_with_deciding_evidence(self):
payload = {
"history": [
{
"step": 1,
"task": "Analyze DGA for T-015",
"server": "fmsr",
"tool": "analyze_dga",
"tool_args": {"transformer_id": "T-015"},
"response": (
'{"diagnosis": "thermal fault", "risk_level": "high", '
'"failure_mode_id": "FM-006"}'
),
"error": None,
"success": True,
}
],
}
adjudication = build_explicit_fault_risk_adjudication(
payload,
enabled=True,
)
self.assertEqual(adjudication["name"], EXPLICIT_FAULT_RISK_ADJUDICATION_NAME)
self.assertEqual(adjudication["decision"], "finalize")
self.assertEqual(adjudication["selected_fault_id"], "FM-006")
self.assertEqual(adjudication["selected_fault_label"], "thermal fault")
self.assertEqual(adjudication["selected_risk_level"], "high")
self.assertEqual(adjudication["deciding_evidence"][0]["tool"], "analyze_dga")
def test_fault_risk_adjudication_refuses_missing_evidence(self):
payload = {
"answer": "Schedule high-priority thermal-fault maintenance.",
"success": True,
"history": [
{
"step": 1,
"task": "Fetch DGA",
"server": "fmsr",
"tool": "get_dga_record",
"tool_args": {"transformer_id": "T-015"},
"response": '{"records": []}',
"error": None,
"success": True,
}
],
}
guarded = apply_explicit_fault_risk_adjudication(payload, enabled=True)
self.assertFalse(guarded["success"])
self.assertEqual(
guarded["fault_risk_adjudication"]["decision"],
"refuse_due_missing_evidence",
)
self.assertIn("Fault/risk adjudication refused", guarded["answer"])
self.assertEqual(
guarded["failed_steps"][-1]["tool"],
EXPLICIT_FAULT_RISK_ADJUDICATION_NAME,
)
def test_fault_risk_adjudication_skips_iot_only_monitoring_task(self):
payload = {
"question": (
"Retrieve load current readings for T-005, compare them to the "
"recent baseline, and report the peak value."
),
"answer": "Peak load current was 520 A at 2024-01-07T12:00:00.",
"success": True,
"scenario": {
"type": "IoT",
"domain_tags": ["IoT"],
"text": "T-005 has elevated load current readings.",
},
"history": [
{
"step": 1,
"task": "Read load current",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-005",
"sensor_id": "load_current_a",
},
"response": '{"readings": [{"value": 520.0}]}',
"error": None,
"success": True,
},
{
"step": 2,
"task": "Check anomalous load-current readings",
"server": "tsfm",
"tool": "detect_anomalies",
"tool_args": {
"transformer_id": "T-005",
"sensor_id": "load_current_a",
},
"response": '{"anomaly_count": 3, "anomaly_rate_pct": 2.5}',
"error": None,
"success": True,
},
],
}
guarded = apply_explicit_fault_risk_adjudication(payload, enabled=True)
self.assertTrue(guarded["success"])
self.assertEqual(
guarded["fault_risk_adjudication"]["decision"],
"not_applicable",
)
self.assertEqual(
guarded["answer"],
"Peak load current was 520 A at 2024-01-07T12:00:00.",
)
self.assertNotIn("failed_steps", guarded)
def test_fault_risk_adjudication_domain_tags_are_exact(self):
monitoring_payload = {
"question": "Retrieve recent load current readings for T-005.",
"scenario": {
"domain_tags": ["wood", "multimedia"],
"text": "Retrieve current readings and report the peak value.",
},
"history": [],
}
work_order_payload = {
"question": "Summarize the current transformer status.",
"scenario": {"domain_tags": ["wo"]},
"history": [],
}
self.assertFalse(_fault_risk_adjudication_applies(monitoring_payload))
self.assertTrue(_fault_risk_adjudication_applies(work_order_payload))
def test_build_fault_risk_adjudication_state_uses_config(self):
with patch.dict(
os.environ,
{
"ENABLE_EXPLICIT_FAULT_RISK_ADJUDICATION": "1",
"ENABLE_MISSING_EVIDENCE_GUARD": "1",
},
):
config = load_fault_risk_adjudication_config()
history = [
{
"step": 1,
"task": "Analyze DGA",
"server": "fmsr",
"tool": "analyze_dga",
"tool_args": {"transformer_id": "T-015"},
"response": '{"diagnosis": "normal", "risk_level": "low"}',
"error": None,
"success": True,
}
]
adjudication = build_fault_risk_adjudication_state(
"Assess T-015 fault risk.",
history,
config,
)
self.assertEqual(adjudication["decision"], "finalize")
self.assertIsNone(fault_risk_adjudication_failed_step(adjudication))
def test_current_missing_evidence_hit_prefers_current_retry_hit(self):
first_miss = {
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": '{"readings": []}',
"error": None,
"success": True,
}
current_miss = {
**first_miss,
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "oil_temp_c",
},
"response": "No readings found for transformer T-015.",
}
hit = current_missing_evidence_hit([first_miss, current_miss], current_miss)
self.assertIsNotNone(hit)
self.assertEqual(hit["target"]["sensor_id"], "oil_temp_c")
self.assertIn("history[1]", hit["source"])
def test_detector_replan_budget_guard_blocks_after_total_attempts_exhausted(self):
with patch.dict(
os.environ,
{
"ENABLE_MISSING_EVIDENCE_REPAIR": "1",
"ENABLE_MISSING_EVIDENCE_GUARD": "1",
"MISSING_EVIDENCE_REPAIR_MAX_ATTEMPTS": "1",
"MISSING_EVIDENCE_REPAIR_MAX_ATTEMPTS_PER_TARGET": "2",
},
):
config = load_missing_evidence_repair_config()
state = build_missing_evidence_repair_state(config)
hit = {
"step": 2,
"tool": "get_sensor_readings",
"target": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"reason": "tool response was empty",
}
record_missing_evidence_retry_attempt(state, hit, action="retry_step")
target_attempts = {repair_target_key(hit): 1}
self.assertFalse(
can_retry_missing_evidence(config, state, hit, target_attempts)
)
def test_repair_state_finalizes_after_successful_retry(self):
with patch.dict(
os.environ,
{
"ENABLE_MISSING_EVIDENCE_REPAIR": "1",
"ENABLE_MISSING_EVIDENCE_GUARD": "1",
},
):
config = load_missing_evidence_repair_config()
state = build_missing_evidence_repair_state(config)
missing_entry = {
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": '{"readings": []}',
"error": None,
"success": True,
}
hit = current_missing_evidence_hit([missing_entry], missing_entry)
self.assertIsNotNone(hit)
state["triggered"] = True
state["attempts"].append(
{
"attempt_index": 1,
"source_step": hit["step"],
"tool": hit["tool"],
"target": hit["target"],
"reason": hit["reason"],
"action": "retry_step",
"result": "scheduled",
}
)
repaired_entry = {
**missing_entry,
"response": '{"readings": [{"timestamp": "2024-01-01T00:00:00", "value": 87.5}]}',
}
finalize_missing_evidence_repair_state(state, [missing_entry, repaired_entry])
self.assertEqual(state["name"], MISSING_EVIDENCE_REPAIR_NAME)
self.assertTrue(state["repaired"])
self.assertEqual(state["final_decision"], "continue")
def test_missing_evidence_guard_allows_evidence_limited_answer(self):
payload = {
"answer": "Cannot determine the maintenance decision because sensor evidence is missing.",
"success": True,
"history": [
{
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {"sensor_id": "winding_temp_top_c"},
"response": "No readings found for transformer T-015.",
"error": None,
"success": True,
}
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertTrue(guarded["success"])
self.assertTrue(guarded["mitigation_guard"]["triggered"])
self.assertFalse(guarded["mitigation_guard"]["blocked_final_answer"])
self.assertTrue(guarded["mitigation_guard"]["evidence_limited_answer"])
def test_missing_evidence_guard_allows_successful_retry_before_answer(self):
payload = {
"answer": "Create the work order using the repaired winding temperature evidence.",
"success": True,
"failed_steps": [],
"history": [
{
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": "No readings found for transformer T-015.",
"error": None,
"success": True,
},
{
"step": 1,
"task": "Retry sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": '{"readings": [{"timestamp": "2024-01-01T00:00:00", "value": 87.5}]}',
"error": None,
"success": True,
},
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertTrue(guarded["success"])
self.assertFalse(guarded["mitigation_guard"]["triggered"])
self.assertEqual(guarded["failed_steps"], [])
def test_missing_evidence_guard_allows_same_step_retry_with_repaired_args(self):
payload = {
"answer": "Use FM-006 sensor correlations for ongoing monitoring.",
"success": True,
"failed_steps": [],
"history": [
{
"step": 4,
"task": "Fetch sensor correlations",
"server": "fmsr",
"tool": "get_sensor_correlation",
"tool_args": {
"failure_mode_id": "Low-temperature overheating",
},
"response": (
'{"error": "Failure mode '
"'Low-temperature overheating' not found.\"}"
),
"error": "Failure mode 'Low-temperature overheating' not found.",
"success": False,
},
{
"step": 4,
"task": "Retry sensor correlations",
"server": "fmsr",
"tool": "get_sensor_correlation",
"tool_args": {"failure_mode_id": "FM-006"},
"response": (
'{"failure_mode_id": "FM-006", ' '"key_gases": ["C2H2", "H2"]}'
),
"error": None,
"success": True,
},
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertTrue(guarded["success"])
self.assertFalse(guarded["mitigation_guard"]["triggered"])
def test_missing_evidence_guard_blocks_work_order_before_retry(self):
payload = {
"answer": "Work order created, then evidence was repaired.",
"success": True,
"history": [
{
"step": 1,
"task": "Fetch sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": "No readings found for transformer T-015.",
"error": None,
"success": True,
},
{
"step": 2,
"task": "Create work order",
"server": "wo",
"tool": "create_work_order",
"tool_args": {"transformer_id": "T-015"},
"response": '{"work_order_id": "WO-123"}',
"error": None,
"success": True,
},
{
"step": 1,
"task": "Retry sensor readings",
"server": "iot",
"tool": "get_sensor_readings",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": '{"readings": [{"timestamp": "2024-01-01T00:00:00", "value": 87.5}]}',
"error": None,
"success": True,
},
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertFalse(guarded["success"])
self.assertTrue(guarded["mitigation_guard"]["blocked_work_order"])
def test_missing_evidence_guard_allows_valid_no_anomaly_result(self):
payload = {
"answer": "No anomalies were found, so defer outage and monitor.",
"success": True,
"history": [
{
"step": 1,
"task": "Detect anomalies",
"server": "tsfm",
"tool": "detect_anomalies",
"tool_args": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"response": (
'{"transformer_id": "T-015", "sensor_id": "winding_temp_top_c", '
'"total_readings": 720, "anomaly_count": 0, "anomalies": []}'
),
"error": None,
"success": True,
}
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertTrue(guarded["success"])
self.assertFalse(guarded["mitigation_guard"]["triggered"])
def test_missing_evidence_guard_scans_get_fault_record(self):
payload = {
"answer": "Use the missing fault record to schedule repair.",
"success": True,
"history": [
{
"step": 1,
"task": "Fetch fault record",
"server": "wo",
"tool": "get_fault_record",
"tool_args": {"fault_id": "F999"},
"response": '{"error": "Fault record F999 not found."}',
"error": None,
"success": True,
}
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertFalse(guarded["success"])
self.assertEqual(
guarded["mitigation_guard"]["hits"][0]["tool"],
"get_fault_record",
)
def test_missing_evidence_guard_scans_aat_tool_calls(self):
payload = {
"answer": "Created a work order for immediate repair.",
"success": True,
"history": [
{
"turn": 1,
"tool_calls": [
{
"name": "get_sensor_readings",
"arguments": {"sensor_id": "oil_temp_c"},
"output": "No data found for transformer T-020.",
},
{
"name": "create_work_order",
"arguments": {"transformer_id": "T-020"},
"output": "WO-123",
},
],
}
],
}
guarded = apply_missing_evidence_final_answer_guard(payload, enabled=True)
self.assertFalse(guarded["success"])
self.assertTrue(guarded["mitigation_guard"]["blocked_work_order"])
def test_missing_evidence_guard_unwraps_aat_text_envelopes(self):
outputs = [
{"type": "text", "text": "[]"},
{"type": "text", "text": '{"readings": []}'},
{"content": [{"type": "text", "text": '{"readings": []}'}]},
]
for output in outputs:
with self.subTest(output=output):
payload = {
"answer": "Schedule immediate maintenance using the empty evidence.",
"success": True,
"history": [
{
"turn": 1,
"tool_calls": [
{
"name": "get_sensor_readings",
"arguments": {
"transformer_id": "T-015",
"sensor_id": "winding_temp_top_c",
},
"output": output,
}
],
}
],
}
guarded = apply_missing_evidence_final_answer_guard(
payload,
enabled=True,
)
self.assertFalse(guarded["success"])
self.assertTrue(guarded["mitigation_guard"]["triggered"])
if __name__ == "__main__":
unittest.main()