Dataset Viewer
Auto-converted to Parquet Duplicate
code
large_stringlengths
193
899
workload_type
large_stringclasses
15 values
gpu_name
large_stringclasses
12 values
gpu_features
large_stringclasses
12 values
runtime_ms
float64
0.01
4.35k
flops
float64
30
21,045B
memory_bytes
int64
160
13.2B
arithmetic_intensity
float64
0
6.37k
dtype_bytes
int64
2
4
workload_params
large_stringlengths
11
79
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.029401
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.043213
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.04236
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.067175
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.042194
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.032627
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.057479
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.060271
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.047451
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.045523
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.057044
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float32, device='cuda') B = torch.randn(64, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.019878
524,288
49,152
10.666667
4
{"M": 64, "N": 64, "K": 64, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.055347
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.043791
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.051259
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.049489
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.045724
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.056497
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.028978
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.03296
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.058059
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.044515
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.061068
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 64) x (64, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 64, dtype=torch.float16, device='cuda') B = torch.randn(64, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.021683
524,288
24,576
21.333333
2
{"M": 64, "N": 64, "K": 64, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.023355
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.041913
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.043544
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.046648
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.056183
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.040023
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.040016
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.053269
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.064568
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.051548
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.022623
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float32, device='cuda') B = torch.randn(256, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.042397
2,097,152
147,456
14.222222
4
{"M": 64, "N": 64, "K": 256, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.062879
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.060706
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.060797
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.064809
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.037732
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.053915
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.029469
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.047584
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.05365
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.024041
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.028216
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 256) x (256, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 256, dtype=torch.float16, device='cuda') B = torch.randn(256, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.057996
2,097,152
73,728
28.444444
2
{"M": 64, "N": 64, "K": 256, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.035004
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.064189
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.031033
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.037495
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.031188
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.046484
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.019261
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.026194
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.050933
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.042775
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.059819
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float32, device='cuda') B = torch.randn(512, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.042135
4,194,304
278,528
15.058824
4
{"M": 64, "N": 64, "K": 512, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.045114
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.064973
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.047348
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.039759
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.03341
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.033329
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.054782
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.038695
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.065029
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.019976
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.037307
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 512) x (512, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 512, dtype=torch.float16, device='cuda') B = torch.randn(512, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.035052
4,194,304
139,264
30.117647
2
{"M": 64, "N": 64, "K": 512, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.071661
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.020933
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.043579
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.046782
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.037203
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.06282
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.033435
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.036598
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.051817
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.046037
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.042846
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float32, device='cuda') B = torch.randn(1024, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.050571
8,388,608
540,672
15.515152
4
{"M": 64, "N": 64, "K": 1024, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.035312
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.052773
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.050574
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.063258
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 80GB
{"gpu_name": "NVIDIA A100 80GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.032275
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L4
{"gpu_name": "NVIDIA L4", "cuda_cores": 7424, "tensor_cores": 232, "memory_gb": 24, "memory_bandwidth_gbps": 300, "base_clock_mhz": 795, "boost_clock_mhz": 2040, "sm_count": 58, "fp32_tflops": 30.3, "fp16_tflops": 121, "tdp_watts": 72, "compute_capability": 8.9, "l2_cache_mb": 48}
0.025499
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA L40S
{"gpu_name": "NVIDIA L40S", "cuda_cores": 18176, "tensor_cores": 568, "memory_gb": 48, "memory_bandwidth_gbps": 864, "base_clock_mhz": 1110, "boost_clock_mhz": 2520, "sm_count": 142, "fp32_tflops": 91.6, "fp16_tflops": 183.2, "tdp_watts": 350, "compute_capability": 8.9, "l2_cache_mb": 96}
0.034444
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 3090
{"gpu_name": "NVIDIA RTX 3090", "cuda_cores": 10496, "tensor_cores": 328, "memory_gb": 24, "memory_bandwidth_gbps": 936, "base_clock_mhz": 1395, "boost_clock_mhz": 1695, "sm_count": 82, "fp32_tflops": 35.6, "fp16_tflops": 71, "tdp_watts": 350, "compute_capability": 8.6, "l2_cache_mb": 6}
0.043579
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX 4090
{"gpu_name": "NVIDIA RTX 4090", "cuda_cores": 16384, "tensor_cores": 512, "memory_gb": 24, "memory_bandwidth_gbps": 1008, "base_clock_mhz": 2235, "boost_clock_mhz": 2520, "sm_count": 128, "fp32_tflops": 82.6, "fp16_tflops": 165.2, "tdp_watts": 450, "compute_capability": 8.9, "l2_cache_mb": 72}
0.029765
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 SXM
{"gpu_name": "NVIDIA H100 SXM", "cuda_cores": 16896, "tensor_cores": 528, "memory_gb": 80, "memory_bandwidth_gbps": 3350, "base_clock_mhz": 1095, "boost_clock_mhz": 1830, "sm_count": 132, "fp32_tflops": 67, "fp16_tflops": 989, "tdp_watts": 700, "compute_capability": 9.0, "l2_cache_mb": 50}
0.033263
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA H100 PCIe
{"gpu_name": "NVIDIA H100 PCIe", "cuda_cores": 14592, "tensor_cores": 456, "memory_gb": 80, "memory_bandwidth_gbps": 2039, "base_clock_mhz": 1095, "boost_clock_mhz": 1620, "sm_count": 114, "fp32_tflops": 48, "fp16_tflops": 756, "tdp_watts": 350, "compute_capability": 9.0, "l2_cache_mb": 50}
0.042784
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 1024) x (1024, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 1024, dtype=torch.float16, device='cuda') B = torch.randn(1024, 64, dtype=torch.float16, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA RTX A6000
{"gpu_name": "NVIDIA RTX A6000", "cuda_cores": 10752, "tensor_cores": 336, "memory_gb": 48, "memory_bandwidth_gbps": 768, "base_clock_mhz": 1410, "boost_clock_mhz": 1860, "sm_count": 84, "fp32_tflops": 38.7, "fp16_tflops": 77.4, "tdp_watts": 300, "compute_capability": 8.6, "l2_cache_mb": 6}
0.055862
8,388,608
270,336
31.030303
2
{"M": 64, "N": 64, "K": 1024, "dtype": "float16"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 2048, dtype=torch.float32, device='cuda') B = torch.randn(2048, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA T4
{"gpu_name": "NVIDIA T4", "cuda_cores": 2560, "tensor_cores": 320, "memory_gb": 16, "memory_bandwidth_gbps": 320, "base_clock_mhz": 585, "boost_clock_mhz": 1590, "sm_count": 40, "fp32_tflops": 8.1, "fp16_tflops": 65, "tdp_watts": 70, "compute_capability": 7.5, "l2_cache_mb": 4}
0.036132
16,777,216
1,064,960
15.753846
4
{"M": 64, "N": 64, "K": 2048, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 2048, dtype=torch.float32, device='cuda') B = torch.randn(2048, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA V100
{"gpu_name": "NVIDIA V100", "cuda_cores": 5120, "tensor_cores": 640, "memory_gb": 32, "memory_bandwidth_gbps": 900, "base_clock_mhz": 1230, "boost_clock_mhz": 1530, "sm_count": 80, "fp32_tflops": 15.7, "fp16_tflops": 125, "tdp_watts": 300, "compute_capability": 7.0, "l2_cache_mb": 6}
0.023584
16,777,216
1,064,960
15.753846
4
{"M": 64, "N": 64, "K": 2048, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 2048, dtype=torch.float32, device='cuda') B = torch.randn(2048, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A10G
{"gpu_name": "NVIDIA A10G", "cuda_cores": 9216, "tensor_cores": 288, "memory_gb": 24, "memory_bandwidth_gbps": 600, "base_clock_mhz": 885, "boost_clock_mhz": 1695, "sm_count": 80, "fp32_tflops": 31.2, "fp16_tflops": 62.5, "tdp_watts": 150, "compute_capability": 8.6, "l2_cache_mb": 6}
0.048067
16,777,216
1,064,960
15.753846
4
{"M": 64, "N": 64, "K": 2048, "dtype": "float32"}
import torch def matmul_kernel(A, B): # Matrix multiplication: (64, 2048) x (2048, 64) -> (64, 64) C = torch.matmul(A, B) return C A = torch.randn(64, 2048, dtype=torch.float32, device='cuda') B = torch.randn(2048, 64, dtype=torch.float32, device='cuda') C = matmul_kernel(A, B) torch.cuda.synchronize()
matmul
NVIDIA A100 40GB
{"gpu_name": "NVIDIA A100 40GB", "cuda_cores": 6912, "tensor_cores": 432, "memory_gb": 40, "memory_bandwidth_gbps": 1555, "base_clock_mhz": 765, "boost_clock_mhz": 1410, "sm_count": 108, "fp32_tflops": 19.5, "fp16_tflops": 312, "tdp_watts": 400, "compute_capability": 8.0, "l2_cache_mb": 40}
0.052525
16,777,216
1,064,960
15.753846
4
{"M": 64, "N": 64, "K": 2048, "dtype": "float32"}
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
23

Models trained or fine-tuned on RajBhope/gpu-runtime-prediction-dataset

Space using RajBhope/gpu-runtime-prediction-dataset 1