File size: 5,913 Bytes
26f2cbf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 | import os
import json
import torch
import matplotlib.pyplot as plt
import torchvision.models as models
from torch.utils.data import DataLoader
from tqdm import tqdm
from verl.cifar10_experiments.sampling_based_rl_objective_experiments import HFImageNet, calculate_loss
from datasets import load_dataset
from torchvision import transforms
def gradient_snr(gradients: torch.Tensor, eps: float = 1e-8) -> torch.Tensor:
"""
Compute the gradient SNR given a set of per-sample gradients.
Args:
gradients: (N, D) tensor of N gradient vectors, each of dimension D.
eps: small constant for numerical stability.
Returns:
snr: scalar tensor, ||mean||^2 / ||var||.
"""
# gradients: (N, D)
mu = gradients.mean(dim=0) # (D,)
var = gradients.var(dim=0, unbiased=False) # (D,)
mean_sq_norm = mu.pow(2).sum() # ||mu||^2
var_sum = var.sum() # ||var||
snr = mean_sq_norm / (var_sum + eps)
return snr, mean_sq_norm, var_sum
CHECKPOINT_PATH = "/data/imagenet256_checkpoints/cross_entropy_1024/checkpoint_latest_final.pth"
BATCH_SIZE = 1024
BATCH_CACHE_PATH = f"batch_{BATCH_SIZE}.pt"
S = 64 # number of gradient samples used to estimate SNR
ADVANTAGE_TYPES = ["grpo", "reinforce_with_baseline", "reinforce_with_p_normalization",]
# ADVANTAGE_TYPES = ["grpo"]
# ROLLOUT_COUNTS = [4, 8, 16, 32, 64]
# ROLLOUT_COUNTS = [64, 128, 256]
# ROLLOUT_COUNTS = [4, 16, 64, 256, 1024, 4096]
ROLLOUT_COUNTS = [2**i for i in range(2, 22, 2)]
print(ROLLOUT_COUNTS)
# ROLLOUT_COUNTS = [2**14, 2**16, 2**18, 2**20]
# ROLLOUT_COUNTS = [2**22, 2**24, 2**26]
SNR_STATS_PATH = "snr_stats_large.json"
MINI_BATCH_SIZE = 512 # number of images per minibatch to avoid OOM
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
# Load model
model = models.resnet50(num_classes=1000)
checkpoint = torch.load(CHECKPOINT_PATH, map_location="cpu")
model.load_state_dict(checkpoint["model_state"])
model = model.to(device)
model.eval()
# Load one batch of BATCH_SIZE images from imagenet (cached)
if os.path.exists(BATCH_CACHE_PATH):
print(f"Loading cached batch from {BATCH_CACHE_PATH}")
batch = torch.load(BATCH_CACHE_PATH)
else:
print("Cache not found, loading dataset...")
mean = (0.485, 0.456, 0.406)
std = (0.229, 0.224, 0.225)
test_transform = transforms.Compose([
transforms.Resize(256),
transforms.CenterCrop(224),
transforms.ToTensor(),
transforms.Normalize(mean=mean, std=std),
])
hf_ds = load_dataset("benjamin-paine/imagenet-1k-256x256")
dataset = HFImageNet(split="validation", transform=test_transform, hf_ds=hf_ds)
loader = DataLoader(dataset, batch_size=BATCH_SIZE, shuffle=False, num_workers=4, pin_memory=True)
batch = next(iter(loader))
torch.save(batch, BATCH_CACHE_PATH)
print(f"Batch cached to {BATCH_CACHE_PATH}")
inputs, targets = batch
inputs, targets = inputs.to(device), targets.to(device)
snr_stats = {}
for advantage_type in ADVANTAGE_TYPES:
snr_stats[advantage_type] = {}
for N in ROLLOUT_COUNTS:
print(f"\n[advantage={advantage_type}, N={N}] collecting {S} gradient samples...")
try:
all_grads = []
num_minibatches = BATCH_SIZE // MINI_BATCH_SIZE
for s in tqdm(range(S)):
model.zero_grad()
for mb_start in range(0, BATCH_SIZE, MINI_BATCH_SIZE):
mb_inputs = inputs [mb_start : mb_start + MINI_BATCH_SIZE]
mb_targets = targets[mb_start : mb_start + MINI_BATCH_SIZE]
logits = model(mb_inputs)
loss = calculate_loss(
logits=logits,
targets=mb_targets,
num_train_rollouts_per_example=N,
advantage_type=advantage_type,
max_k=None,
)
# scale so accumulated gradient == average over full batch
(loss / num_minibatches).backward()
flat_grad = torch.cat([
p.grad.detach().view(-1)
for p in model.parameters()
if p.grad is not None
])
all_grads.append(flat_grad)
gradients = torch.stack(all_grads) # (S, D)
snr, mean_sq_norm, var_sum = gradient_snr(gradients)
snr, mean_sq_norm, var_sum = snr.item(), mean_sq_norm.item(), var_sum.item()
print(f" SNR = {snr:.6f}, mean = {mean_sq_norm:.6f}, var = {var_sum:.6f}")
snr_stats[advantage_type][N] = {
"snr": snr,
"mean": mean_sq_norm,
"var": var_sum,
}
except Exception as e:
print(f" ERROR: {e}")
snr_stats[advantage_type][N] = None
with open(SNR_STATS_PATH, "w") as f:
json.dump(snr_stats, f, indent=2)
print(f"\nSaved SNR stats to {SNR_STATS_PATH}")
# --- Plot ---
LABEL_MAP = {
"reinforce_with_baseline": "RLOO",
"grpo": "GRPO",
"reinforce_with_p_normalization": "MaxRL",
}
with open(SNR_STATS_PATH) as f:
plot_data = json.load(f)
fig, ax = plt.subplots(figsize=(8, 5))
for advantage_type, rollout_snrs in plot_data.items():
xs = [int(k) for k in rollout_snrs]
ys = [v["snr"] if isinstance(v, dict) else v for v in rollout_snrs.values()]
label = LABEL_MAP.get(advantage_type, advantage_type)
ax.plot(xs, ys, marker="o", label=label)
ax.set_xscale("log", base=2)
ax.set_xlabel("Rollouts (N)")
ax.set_ylabel("Gradient SNR")
ax.set_title(f"Gradient SNR vs Rollouts (S={S} gradient samples)")
ax.legend()
ax.grid(True, which="both", linestyle="--", alpha=0.5)
plt.tight_layout()
plt.savefig("snr.png", dpi=150)
print("Saved plot to snr.png") |