Infatoshi's picture
initial upload: 7 problem definitions
80692f2 verified
name: 04_kahan_softmax
display_name: "Kahan-corrected Softmax"
precision: fp32
regime: memory # softmax is bandwidth-bound: 2 passes over the input tensor
# Softmax FLOPs: per-element exp + 2 reductions + divide. Roughly 5 flops/elt.
flops_formula: "5 * batch * vocab"
# Bytes moved: read x once, write y once. Both fp32.
bytes_formula: "batch * vocab * 4 + batch * vocab * 4"
hardware: [RTX_PRO_6000]
peak_tflops_key: fp32
peak_bandwidth_key: dram
# TIGHTER than default (fp32 default is 1e-4). This problem exists
# specifically to test whether the agent uses compensated summation, so
# we squeeze the tolerance to 1e-5 — naive fp16 sum across 256K elements
# drifts past this; fp32 accumulation passes; Kahan/fp32 always passes.
tolerance:
"torch.float32": {"atol": 1.0e-5, "rtol": 1.0e-5}
# Forbidden ops — block the obvious "just call the library" cheats. The
# agent must implement softmax themselves with explicit (compensated)
# summation logic.
forbidden:
- "torch.nn.functional.softmax"
- "torch.softmax"
- "F.softmax"
- "liger_kernel.softmax"
- "liger_kernel.transformers.softmax"
- ".softmax("
sota:
name: "Liger-Kernel Softmax (Triton)"
url: "https://github.com/linkedin/Liger-Kernel"
function: "liger_kernel.ops.softmax.LigerSoftmaxFunction"
deps:
- "liger-kernel>=0.5.0"
reference_throughput_gbps_h100: 2800
num_correct_trials: 3
num_perf_trials: 30