RajBhope/gpu-runtime-predictor
Tabular Regression • Updated
code large_stringlengths 193 899 | workload_type large_stringclasses 15
values | gpu_name large_stringclasses 12
values | gpu_features large_stringclasses 12
values | runtime_ms float64 0.01 4.35k | flops float64 30 21,045B | memory_bytes int64 160 13.2B | arithmetic_intensity float64 0 6.37k | dtype_bytes int64 2 4 | workload_params large_stringlengths 11 79 |
|---|---|---|---|---|---|---|---|---|---|
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.029401 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.043213 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.04236 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.067175 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.042194 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.032627 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.057479 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.060271 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.047451 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.045523 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.057044 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float32, device='cuda')
B = torch.randn(64, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.019878 | 524,288 | 49,152 | 10.666667 | 4 | {"M": 64, "N": 64, "K": 64, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.055347 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.043791 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.051259 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.049489 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.045724 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.056497 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.028978 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.03296 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.058059 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.044515 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.061068 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 64) x (64, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 64, dtype=torch.float16, device='cuda')
B = torch.randn(64, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.021683 | 524,288 | 24,576 | 21.333333 | 2 | {"M": 64, "N": 64, "K": 64, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.023355 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.041913 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.043544 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.046648 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.056183 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.040023 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.040016 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.053269 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.064568 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.051548 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.022623 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float32, device='cuda')
B = torch.randn(256, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.042397 | 2,097,152 | 147,456 | 14.222222 | 4 | {"M": 64, "N": 64, "K": 256, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.062879 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.060706 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.060797 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.064809 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.037732 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.053915 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.029469 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.047584 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.05365 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.024041 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.028216 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 256) x (256, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 256, dtype=torch.float16, device='cuda')
B = torch.randn(256, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.057996 | 2,097,152 | 73,728 | 28.444444 | 2 | {"M": 64, "N": 64, "K": 256, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.035004 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.064189 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.031033 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.037495 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.031188 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.046484 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.019261 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.026194 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.050933 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.042775 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.059819 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float32, device='cuda')
B = torch.randn(512, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.042135 | 4,194,304 | 278,528 | 15.058824 | 4 | {"M": 64, "N": 64, "K": 512, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.045114 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.064973 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.047348 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.039759 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.03341 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.033329 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.054782 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.038695 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.065029 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.019976 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.037307 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 512) x (512, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 512, dtype=torch.float16, device='cuda')
B = torch.randn(512, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.035052 | 4,194,304 | 139,264 | 30.117647 | 2 | {"M": 64, "N": 64, "K": 512, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.071661 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.020933 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.043579 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.046782 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.037203 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.06282 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.033435 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.036598 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.051817 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.046037 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.042846 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float32, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.050571 | 8,388,608 | 540,672 | 15.515152 | 4 | {"M": 64, "N": 64, "K": 1024, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.035312 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.052773 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.050574 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.063258 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 80GB | {"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.032275 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L4 | {"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48} | 0.025499 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA L40S | {"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96} | 0.034444 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 3090 | {"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.043579 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX 4090 | {"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72} | 0.029765 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 SXM | {"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.033263 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA H100 PCIe | {"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50} | 0.042784 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 1024, dtype=torch.float16, device='cuda')
B = torch.randn(1024, 64, dtype=torch.float16, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA RTX A6000 | {"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.055862 | 8,388,608 | 270,336 | 31.030303 | 2 | {"M": 64, "N": 64, "K": 1024, "dtype": "float16"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 2048, dtype=torch.float32, device='cuda')
B = torch.randn(2048, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA T4 | {"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4} | 0.036132 | 16,777,216 | 1,064,960 | 15.753846 | 4 | {"M": 64, "N": 64, "K": 2048, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 2048, dtype=torch.float32, device='cuda')
B = torch.randn(2048, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA V100 | {"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6} | 0.023584 | 16,777,216 | 1,064,960 | 15.753846 | 4 | {"M": 64, "N": 64, "K": 2048, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 2048, dtype=torch.float32, device='cuda')
B = torch.randn(2048, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A10G | {"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6} | 0.048067 | 16,777,216 | 1,064,960 | 15.753846 | 4 | {"M": 64, "N": 64, "K": 2048, "dtype": "float32"} |
import torch
def matmul_kernel(A, B):
# Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64)
C = torch.matmul(A, B)
return C
A = torch.randn(64, 2048, dtype=torch.float32, device='cuda')
B = torch.randn(2048, 64, dtype=torch.float32, device='cuda')
C = matmul_kernel(A, B)
torch.cuda.synchronize()
| matmul | NVIDIA A100 40GB | {"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40} | 0.052525 | 16,777,216 | 1,064,960 | 15.753846 | 4 | {"M": 64, "N": 64, "K": 2048, "dtype": "float32"} |