Infatoshi's picture
initial upload: 7 problem definitions
80692f2 verified
"""Naive W4A16 weight-only quantized GEMM reference (correctness only).
AWQ/GPTQ-style scheme:
x: (M, K) bf16
w_q: (K // 2, N) uint8 -- two int4 weights packed per byte (low nibble = even-K, high = odd-K)
scales: (K // group, N) bf16
zeros: (K // group, N) bf16 -- asymmetric (stored already as float zero-point)
out: (M, N) bf16
Dequant (per group along K):
w_bf[k, n] = (w_q[k, n] - zeros[k // group, n]) * scales[k // group, n]
where w_q[k, n] is the unpacked 4-bit value (0..15).
This reference unpacks to a full bf16 matrix and then runs torch.matmul. Slow and
memory-heavy on the dequant; the agent's solution must fuse unpack+GEMM.
"""
from __future__ import annotations
import torch
import torch.nn as nn
OP_TYPE = "gemm_w4a16"
SUPPORTED_PRECISIONS = ["int4_bf16"]
HARDWARE_REQUIRED = ["RTX_PRO_6000", "H100", "B200"]
GROUP_SIZE = 128
def _pack_int4(w_q: torch.Tensor) -> torch.Tensor:
"""Pack (K, N) uint8 in [0,15] into (K//2, N) uint8.
Even rows go in the low nibble, odd rows in the high nibble.
"""
K, N = w_q.shape
assert K % 2 == 0
lo = w_q[0::2].to(torch.uint8) & 0xF
hi = w_q[1::2].to(torch.uint8) & 0xF
return (lo | (hi << 4)).contiguous()
def _unpack_int4(w_packed: torch.Tensor, K: int) -> torch.Tensor:
"""Unpack (K//2, N) uint8 -> (K, N) uint8 in [0,15]."""
Kh, N = w_packed.shape
assert Kh * 2 == K
out = torch.empty((K, N), dtype=torch.uint8, device=w_packed.device)
out[0::2] = w_packed & 0xF
out[1::2] = (w_packed >> 4) & 0xF
return out
class Model(nn.Module):
"""W4A16 GEMM: y = x @ dequant(w_q, scales, zeros).
Buffers are registered (not Parameters) so state_dict carries them across to
the agent's solution. Initialization picks scales/zeros from a normal weight,
then quantizes deterministically.
"""
def __init__(self, M: int, N: int, K: int, group_size: int = GROUP_SIZE):
super().__init__()
assert K % group_size == 0, "K must be divisible by group_size"
assert K % 2 == 0, "K must be even (int4 packing)"
self.M, self.N, self.K = M, N, K
self.group_size = group_size
n_groups = K // group_size
# Synthetic quant: take a random bf16 weight, compute per-group asym
# quant params, then pack. This produces a *correct* set of (w_q, s, z)
# triples that round-trip cleanly under the dequant formula.
torch.manual_seed(0xC0DE ^ (M * 1315423911 + N * 2654435761 + K))
w_full = torch.randn(K, N, dtype=torch.float32) * 0.02
w_g = w_full.view(n_groups, group_size, N)
w_min = w_g.min(dim=1, keepdim=True).values # (n_groups, 1, N)
w_max = w_g.max(dim=1, keepdim=True).values
scales = (w_max - w_min).clamp_min(1e-8) / 15.0 # (n_groups, 1, N)
zeros = (-w_min / scales).round().clamp(0, 15) # (n_groups, 1, N)
# Quantize
w_q = ((w_g / scales) + zeros).round().clamp(0, 15).to(torch.uint8)
w_q = w_q.view(K, N)
scales_2d = scales.squeeze(1).to(torch.bfloat16) # (n_groups, N)
zeros_2d = zeros.squeeze(1).to(torch.bfloat16) # (n_groups, N)
w_packed = _pack_int4(w_q) # (K//2, N)
self.register_buffer("w_q", w_packed) # uint8
self.register_buffer("scales", scales_2d) # bf16
self.register_buffer("zeros", zeros_2d) # bf16
def forward(self, x: torch.Tensor) -> torch.Tensor:
# Naive: unpack -> dequant -> matmul.
K = self.K
w_unpacked = _unpack_int4(self.w_q, K).to(torch.bfloat16) # (K, N) in [0,15]
# Broadcast scales/zeros along the group axis.
scales = self.scales.repeat_interleave(self.group_size, dim=0) # (K, N) bf16
zeros = self.zeros.repeat_interleave(self.group_size, dim=0) # (K, N) bf16
w_bf = (w_unpacked - zeros) * scales # (K, N) bf16
return x.to(torch.bfloat16) @ w_bf # (M, N) bf16
M = 1
N = 12288
K = 4096
def get_inputs():
x = torch.randn(M, K, dtype=torch.bfloat16)
return [x]
def get_init_inputs():
return [M, N, K]