Infatoshi's picture
initial upload: 7 problem definitions
80692f2 verified
"""Correctness runner for TopK.
Runs solution.Model vs reference.Model across all shapes in shapes.py, 3 seeds
each. Top-k correctness has two parts:
1. VALUES: sol_values must match ref_values within fp32 tol. Both are
returned sorted descending, so positional comparison is well-defined.
2. INDICES: lenient — we do NOT require sol_indices == ref_indices because
ties in x can yield multiple valid index sets. Instead we gather x at
sol_indices and check those values match ref_values within tol. This
catches "wrong indices" without false-failing on legitimate tie-breaks.
Also rejects forbidden ops by grep.
"""
import re
import sys
from pathlib import Path
import torch
import yaml
REPO_ROOT = Path(__file__).resolve().parents[2]
sys.path.insert(0, str(REPO_ROOT))
from src.eval.correctness import check_correctness # noqa: E402
def main():
try:
import reference
import shapes
import solution
except Exception as e:
print(f"FAIL: import error: {e}")
sys.exit(1)
problem_yaml = Path("problem.yaml")
meta = yaml.safe_load(problem_yaml.read_text()) if problem_yaml.exists() else {}
# --- Forbidden-op check ------------------------------------------------
sol_src = Path("solution.py").read_text() if Path("solution.py").exists() else ""
for forbidden in meta.get("forbidden", []):
pat = re.escape(forbidden)
if re.search(pat, sol_src):
print(f"FAIL: forbidden op used: {forbidden}")
sys.exit(1)
device = torch.device("cuda:0")
tol_override = meta.get("tolerance") or None
all_shapes = shapes.SHAPES
for shape_idx, shape in enumerate(all_shapes):
reference.batch = shape["batch"]
reference.n = shape["n"]
reference.k = shape["k"]
init_args = reference.get_init_inputs()
ref_model = reference.Model(*init_args).to(device).eval()
sol_model = solution.Model(*init_args).to(device).eval()
sd = ref_model.state_dict()
try:
sol_model.load_state_dict(sd, strict=True)
except RuntimeError as e:
print(f"FAIL: state_dict mismatch at shape {shape_idx} ({shape}): {e}")
sys.exit(1)
for seed in (42, 123, 456):
torch.manual_seed(seed)
torch.cuda.manual_seed_all(seed)
inputs = [t.to(device) for t in reference.get_inputs()]
with torch.no_grad():
ref_values, ref_indices = ref_model(*inputs)
sol_out = sol_model(*inputs)
if not (isinstance(sol_out, (tuple, list)) and len(sol_out) == 2):
print(f"FAIL: shape {shape_idx} {shape} seed {seed}: "
f"solution must return (values, indices); got {type(sol_out)}")
sys.exit(1)
sol_values, sol_indices = sol_out
# Shape checks
expected_shape = (shape["batch"], shape["k"])
if tuple(sol_values.shape) != expected_shape:
print(f"FAIL: shape {shape_idx} values shape {tuple(sol_values.shape)} "
f"!= expected {expected_shape}")
sys.exit(1)
if tuple(sol_indices.shape) != expected_shape:
print(f"FAIL: shape {shape_idx} indices shape {tuple(sol_indices.shape)} "
f"!= expected {expected_shape}")
sys.exit(1)
# 1. Strict-ish values check (positional, both are sorted desc)
ok, msg = check_correctness(
ref_values.float(), sol_values.float(),
dtype=torch.float32,
override=tol_override,
)
if not ok:
print(f"FAIL: shape {shape_idx} {shape} seed {seed} values: {msg}")
sys.exit(1)
# 2. Lenient indices check: gather x at sol_indices, compare to ref_values.
# This handles ties without false negatives.
x = inputs[0]
sol_idx_long = sol_indices.to(torch.int64)
if sol_idx_long.min() < 0 or sol_idx_long.max() >= shape["n"]:
print(f"FAIL: shape {shape_idx} indices out of range "
f"[{int(sol_idx_long.min())}, {int(sol_idx_long.max())}]")
sys.exit(1)
gathered = torch.gather(x, dim=-1, index=sol_idx_long)
ok, msg = check_correctness(
ref_values.float(), gathered.float(),
dtype=torch.float32,
override=tol_override,
)
if not ok:
print(f"FAIL: shape {shape_idx} {shape} seed {seed} indices "
f"(gather mismatch): {msg}")
sys.exit(1)
_emit_framework_label()
print("PASS")
def _emit_framework_label():
patterns = [
("ptx", r"asm\s+volatile|asm\s*\(|mma\.sync|tcgen05\."),
("cutlass3", r"\bcute::|cutlass/gemm/collective|cutlass::arch::Sm(9|10|12)"),
("cutlass2", r"cutlass/gemm/device/gemm|cutlass::gemm::device"),
("cuda_wmma", r"\bnvcuda::wmma\b|wmma::fragment"),
("triton", r"import\s+triton\b|@triton\.jit|\btl\.dot\b"),
("cuda_raw", r"torch\.utils\.cpp_extension\.load_inline|__global__\s+void"),
]
sol = Path("solution.py")
if not sol.exists():
return
code = sol.read_text()
label = "unknown"
for name, pat in patterns:
if re.search(pat, code):
label = name
break
Path("framework.txt").write_text(label + "\n")
if __name__ == "__main__":
main()