| """Roofline benchmark for Kahan-corrected softmax. |
| |
| For each shape: times eager reference, compiled reference, SOTA (if |
| available), and the agent's solution. Reports achieved TFLOPS, GB/s, and |
| peak_fraction. Softmax is memory-bound, so the score is GB/s / peak_dram. |
| |
| Output lines the harness picks up: |
| shape=<idx> variant=<name> tflops=<N> gbps=<N> ms=<N> |
| peak_fraction: <N> (geomean over shapes of solution's peak_fraction) |
| """ |
| import sys |
| from math import exp, log |
| from pathlib import Path |
|
|
| import torch |
| import yaml |
|
|
| REPO_ROOT = Path(__file__).resolve().parents[2] |
| sys.path.insert(0, str(REPO_ROOT)) |
|
|
| from src.eval.roofline import compute_gbps, compute_tflops, peak_fraction |
| from src.eval.timing import time_fn |
| from src.hardware import get as get_hw |
|
|
|
|
| def _eval_formula(expr: str, vars: dict) -> float: |
| return float(eval(expr, {"__builtins__": {}}, vars)) |
|
|
|
|
| def _make_inputs(batch: int, vocab: int, extreme: bool) -> torch.Tensor: |
| if extreme: |
| x = torch.randn(batch, vocab) * 2.0 |
| idx = torch.randint(0, vocab, (batch, 4)) |
| x.scatter_(1, idx, 30.0) |
| else: |
| x = torch.randn(batch, vocab) * 4.0 |
| return x.to(torch.float32) |
|
|
|
|
| def main(): |
| import reference |
| import shapes |
| import solution |
|
|
| meta = yaml.safe_load(Path("problem.yaml").read_text()) |
| hw = get_hw(meta["hardware"][0]) |
| peak_tflops = hw.peak_tflops_dense.get(meta["peak_tflops_key"], 0.0) |
| peak_gbps = hw.peak_bandwidth_gb_s |
| regime = meta.get("regime", "memory") |
| flops_formula = meta["flops_formula"] |
| bytes_formula = meta["bytes_formula"] |
| num_perf_trials = int(meta.get("num_perf_trials", 30)) |
|
|
| device = torch.device("cuda:0") |
|
|
| try: |
| import sota as sota_mod |
| has_sota = sota_mod.is_available() |
| except Exception: |
| has_sota = False |
|
|
| sol_fractions: list[float] = [] |
|
|
| for shape_idx, shape in enumerate(shapes.SHAPES): |
| batch = shape["batch"] |
| vocab = shape["vocab"] |
| extreme = shape.get("extreme", False) |
|
|
| reference.BATCH = batch |
| reference.VOCAB = vocab |
|
|
| init_args = reference.get_init_inputs() |
| ref_model = reference.Model(*init_args).to(device).eval() |
| sol_model = solution.Model(*init_args).to(device).eval() |
| sd = ref_model.state_dict() |
| try: |
| sol_model.load_state_dict(sd, strict=True) |
| except RuntimeError: |
| pass |
|
|
| torch.manual_seed(2026) |
| x = _make_inputs(batch, vocab, extreme).to(device) |
| inputs = [x] |
|
|
| flops = _eval_formula(flops_formula, {"batch": batch, "vocab": vocab}) |
| bytes_moved = _eval_formula(bytes_formula, {"batch": batch, "vocab": vocab}) |
|
|
| ms_eager = time_fn(ref_model, inputs, iters=num_perf_trials) |
|
|
| try: |
| comp = torch.compile(ref_model, mode="reduce-overhead") |
| ms_comp = time_fn(comp, inputs, iters=num_perf_trials) |
| except Exception as e: |
| print(f" [compile fallback] {type(e).__name__}: {e}") |
| ms_comp = None |
|
|
| ms_sota = None |
| if has_sota: |
| try: |
| def sota_fn(t): |
| return sota_mod.sota_forward(t) |
| ms_sota = time_fn(sota_fn, inputs, iters=num_perf_trials) |
| except Exception as e: |
| print(f" [sota unavailable] {type(e).__name__}: {e}") |
|
|
| ms_sol = time_fn(sol_model, inputs, iters=num_perf_trials) |
|
|
| for variant, ms in [ |
| ("eager", ms_eager), |
| ("compiled", ms_comp), |
| ("sota", ms_sota), |
| ("solution", ms_sol), |
| ]: |
| if ms is None: |
| continue |
| tflops = compute_tflops(flops, ms) |
| gbps = compute_gbps(bytes_moved, ms) |
| print(f"shape={shape_idx} variant={variant} tflops={tflops:.3f} gbps={gbps:.3f} ms={ms:.3f}") |
|
|
| sol_tflops = compute_tflops(flops, ms_sol) |
| sol_gbps = compute_gbps(bytes_moved, ms_sol) |
| if regime == "compute": |
| frac = peak_fraction(sol_tflops, peak_tflops) |
| else: |
| frac = peak_fraction(sol_gbps, peak_gbps) |
| sol_fractions.append(frac) |
| print(f"shape={shape_idx} solution_peak_fraction={frac:.4f}") |
|
|
| gmean = exp(sum(log(max(f, 1e-9)) for f in sol_fractions) / len(sol_fractions)) |
| print(f"peak_fraction: {gmean:.4f}") |
| print(f"RESULT: {'OK' if gmean >= 0.1 else 'LOW'}") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|