guanning commited on
Commit
26f2cbf
·
verified ·
1 Parent(s): ef17d5e

Add files using upload-large-folder tool

Browse files
Files changed (49) hide show
  1. .gitattributes +6 -0
  2. README.md +19 -0
  3. __pycache__/qwen3_variance_analysis_auto.cpython-310.pyc +0 -0
  4. __pycache__/smollm_variance_analysis.cpython-313.pyc +0 -0
  5. damans_code.py +173 -0
  6. data/Maze/variance/0000.jsonl +3 -0
  7. data/Maze/variance/5000.jsonl +3 -0
  8. data/Maze/variance/8000.jsonl +3 -0
  9. data/Qwen3/1.7b_512x512.jsonl +3 -0
  10. data/Qwen3/4b_1024x128.jsonl +3 -0
  11. data/README.md +18 -0
  12. data/SmolLM/512x512.jsonl +3 -0
  13. maze_variance_analysis.py +0 -0
  14. maze_variance_analysis_auto.py +400 -0
  15. outputs/Maze/results.json +58 -0
  16. outputs/Maze/variance_plot.pdf +0 -0
  17. outputs/Maze/variance_plot.png +3 -0
  18. outputs/Qwen3/.gitkeep +1 -0
  19. outputs/SmolLM/results.json +50 -0
  20. outputs/SmolLM/variance_plot.pdf +0 -0
  21. outputs/SmolLM/variance_plot.png +3 -0
  22. outputs/SmolLM_SNR/.gitkeep +1 -0
  23. outputs/old/results_bs16_nr128_nb4_r5_blFalse.json +80 -0
  24. outputs/old/results_bs16_nr128_nb4_r5_blTrue.json +80 -0
  25. outputs/old/results_bs16_nr16_nb4_r5_blFalse.json +80 -0
  26. outputs/old/results_bs16_nr16_nb4_r5_blTrue.json +80 -0
  27. outputs/old/results_bs16_nr32_nb4_r5_blFalse.json +80 -0
  28. outputs/old/results_bs16_nr32_nb4_r5_blTrue.json +80 -0
  29. outputs/old/results_bs16_nr4_nb4_r5_blFalse.json +80 -0
  30. outputs/old/results_bs16_nr4_nb4_r5_blTrue.json +80 -0
  31. outputs/old/results_bs16_nr64_nb4_r5_blFalse.json +80 -0
  32. outputs/old/results_bs16_nr64_nb4_r5_blTrue.json +80 -0
  33. outputs/old/results_bs16_nr8_nb4_r5_blFalse.json +80 -0
  34. outputs/old/results_bs16_nr8_nb4_r5_blTrue.json +80 -0
  35. outputs/variance_comparison_all.pdf +0 -0
  36. outputs/variance_comparison_all.png +3 -0
  37. plot_all_variance.py +81 -0
  38. plot_variance.py +39 -0
  39. qwen3_variance_analysis_auto.py +400 -0
  40. qwen3_variance_analysis_resume.py +75 -0
  41. run_maze_auto.sh +38 -0
  42. run_qwen3_auto.sh +38 -0
  43. run_qwen3_resume.sh +38 -0
  44. run_smollm_auto.sh +38 -0
  45. run_smollm_snr.sh +38 -0
  46. smollm_snr_analysis_auto.py +442 -0
  47. smollm_variance_analysis.py +478 -0
  48. smollm_variance_analysis_auto.py +400 -0
  49. variance_plot.png +3 -0
.gitattributes CHANGED
@@ -58,3 +58,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
58
  # Video files - compressed
59
  *.mp4 filter=lfs diff=lfs merge=lfs -text
60
  *.webm filter=lfs diff=lfs merge=lfs -text
61
+ data/SmolLM/512x512.jsonl filter=lfs diff=lfs merge=lfs -text
62
+ data/Qwen3/1.7b_512x512.jsonl filter=lfs diff=lfs merge=lfs -text
63
+ data/Qwen3/4b_1024x128.jsonl filter=lfs diff=lfs merge=lfs -text
64
+ data/Maze/variance/5000.jsonl filter=lfs diff=lfs merge=lfs -text
65
+ data/Maze/variance/0000.jsonl filter=lfs diff=lfs merge=lfs -text
66
+ data/Maze/variance/8000.jsonl filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pretty_name: Variance Analysis
3
+ ---
4
+
5
+ # Variance Analysis
6
+
7
+ This repository contains the data, scripts, and generated figures used for
8
+ variance analysis experiments.
9
+
10
+ ## Contents
11
+
12
+ - `data/SmolLM`: SmolLM GSM8K rollout data.
13
+ - `data/Qwen3`: Qwen3 math rollout data, including 1.7B `512 x 512` and
14
+ 4B `1024 x 128` samples.
15
+ - `data/Maze/variance`: Maze rollout data for variance analysis.
16
+ - `outputs`: generated JSON summaries and figures.
17
+ - `*.py` and `run_*.sh`: analysis, plotting, and Slurm launch scripts.
18
+
19
+ See `data/README.md` for additional data details and source model links.
__pycache__/qwen3_variance_analysis_auto.cpython-310.pyc ADDED
Binary file (10.9 kB). View file
 
__pycache__/smollm_variance_analysis.cpython-313.pyc ADDED
Binary file (20.3 kB). View file
 
damans_code.py ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import json
3
+ import torch
4
+ import matplotlib.pyplot as plt
5
+ import torchvision.models as models
6
+ from torch.utils.data import DataLoader
7
+ from tqdm import tqdm
8
+
9
+ from verl.cifar10_experiments.sampling_based_rl_objective_experiments import HFImageNet, calculate_loss
10
+
11
+ from datasets import load_dataset
12
+ from torchvision import transforms
13
+
14
+ def gradient_snr(gradients: torch.Tensor, eps: float = 1e-8) -> torch.Tensor:
15
+ """
16
+ Compute the gradient SNR given a set of per-sample gradients.
17
+
18
+ Args:
19
+ gradients: (N, D) tensor of N gradient vectors, each of dimension D.
20
+ eps: small constant for numerical stability.
21
+
22
+ Returns:
23
+ snr: scalar tensor, ||mean||^2 / ||var||.
24
+ """
25
+ # gradients: (N, D)
26
+ mu = gradients.mean(dim=0) # (D,)
27
+ var = gradients.var(dim=0, unbiased=False) # (D,)
28
+ mean_sq_norm = mu.pow(2).sum() # ||mu||^2
29
+ var_sum = var.sum() # ||var||
30
+ snr = mean_sq_norm / (var_sum + eps)
31
+ return snr, mean_sq_norm, var_sum
32
+
33
+
34
+ CHECKPOINT_PATH = "/data/imagenet256_checkpoints/cross_entropy_1024/checkpoint_latest_final.pth"
35
+ BATCH_SIZE = 1024
36
+ BATCH_CACHE_PATH = f"batch_{BATCH_SIZE}.pt"
37
+ S = 64 # number of gradient samples used to estimate SNR
38
+
39
+ ADVANTAGE_TYPES = ["grpo", "reinforce_with_baseline", "reinforce_with_p_normalization",]
40
+ # ADVANTAGE_TYPES = ["grpo"]
41
+ # ROLLOUT_COUNTS = [4, 8, 16, 32, 64]
42
+ # ROLLOUT_COUNTS = [64, 128, 256]
43
+ # ROLLOUT_COUNTS = [4, 16, 64, 256, 1024, 4096]
44
+ ROLLOUT_COUNTS = [2**i for i in range(2, 22, 2)]
45
+ print(ROLLOUT_COUNTS)
46
+ # ROLLOUT_COUNTS = [2**14, 2**16, 2**18, 2**20]
47
+
48
+ # ROLLOUT_COUNTS = [2**22, 2**24, 2**26]
49
+
50
+ SNR_STATS_PATH = "snr_stats_large.json"
51
+ MINI_BATCH_SIZE = 512 # number of images per minibatch to avoid OOM
52
+
53
+ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
54
+
55
+ # Load model
56
+ model = models.resnet50(num_classes=1000)
57
+ checkpoint = torch.load(CHECKPOINT_PATH, map_location="cpu")
58
+ model.load_state_dict(checkpoint["model_state"])
59
+ model = model.to(device)
60
+ model.eval()
61
+
62
+ # Load one batch of BATCH_SIZE images from imagenet (cached)
63
+ if os.path.exists(BATCH_CACHE_PATH):
64
+ print(f"Loading cached batch from {BATCH_CACHE_PATH}")
65
+ batch = torch.load(BATCH_CACHE_PATH)
66
+ else:
67
+ print("Cache not found, loading dataset...")
68
+ mean = (0.485, 0.456, 0.406)
69
+ std = (0.229, 0.224, 0.225)
70
+
71
+ test_transform = transforms.Compose([
72
+ transforms.Resize(256),
73
+ transforms.CenterCrop(224),
74
+ transforms.ToTensor(),
75
+ transforms.Normalize(mean=mean, std=std),
76
+ ])
77
+
78
+ hf_ds = load_dataset("benjamin-paine/imagenet-1k-256x256")
79
+ dataset = HFImageNet(split="validation", transform=test_transform, hf_ds=hf_ds)
80
+ loader = DataLoader(dataset, batch_size=BATCH_SIZE, shuffle=False, num_workers=4, pin_memory=True)
81
+
82
+ batch = next(iter(loader))
83
+ torch.save(batch, BATCH_CACHE_PATH)
84
+ print(f"Batch cached to {BATCH_CACHE_PATH}")
85
+
86
+ inputs, targets = batch
87
+ inputs, targets = inputs.to(device), targets.to(device)
88
+
89
+ snr_stats = {}
90
+
91
+ for advantage_type in ADVANTAGE_TYPES:
92
+ snr_stats[advantage_type] = {}
93
+ for N in ROLLOUT_COUNTS:
94
+ print(f"\n[advantage={advantage_type}, N={N}] collecting {S} gradient samples...")
95
+ try:
96
+ all_grads = []
97
+
98
+ num_minibatches = BATCH_SIZE // MINI_BATCH_SIZE
99
+
100
+ for s in tqdm(range(S)):
101
+ model.zero_grad()
102
+
103
+ for mb_start in range(0, BATCH_SIZE, MINI_BATCH_SIZE):
104
+ mb_inputs = inputs [mb_start : mb_start + MINI_BATCH_SIZE]
105
+ mb_targets = targets[mb_start : mb_start + MINI_BATCH_SIZE]
106
+
107
+ logits = model(mb_inputs)
108
+
109
+ loss = calculate_loss(
110
+ logits=logits,
111
+ targets=mb_targets,
112
+ num_train_rollouts_per_example=N,
113
+ advantage_type=advantage_type,
114
+ max_k=None,
115
+ )
116
+
117
+ # scale so accumulated gradient == average over full batch
118
+ (loss / num_minibatches).backward()
119
+
120
+ flat_grad = torch.cat([
121
+ p.grad.detach().view(-1)
122
+ for p in model.parameters()
123
+ if p.grad is not None
124
+ ])
125
+ all_grads.append(flat_grad)
126
+
127
+ gradients = torch.stack(all_grads) # (S, D)
128
+ snr, mean_sq_norm, var_sum = gradient_snr(gradients)
129
+ snr, mean_sq_norm, var_sum = snr.item(), mean_sq_norm.item(), var_sum.item()
130
+ print(f" SNR = {snr:.6f}, mean = {mean_sq_norm:.6f}, var = {var_sum:.6f}")
131
+ snr_stats[advantage_type][N] = {
132
+ "snr": snr,
133
+ "mean": mean_sq_norm,
134
+ "var": var_sum,
135
+ }
136
+
137
+ except Exception as e:
138
+ print(f" ERROR: {e}")
139
+ snr_stats[advantage_type][N] = None
140
+
141
+ with open(SNR_STATS_PATH, "w") as f:
142
+ json.dump(snr_stats, f, indent=2)
143
+
144
+ print(f"\nSaved SNR stats to {SNR_STATS_PATH}")
145
+
146
+ # --- Plot ---
147
+ LABEL_MAP = {
148
+ "reinforce_with_baseline": "RLOO",
149
+ "grpo": "GRPO",
150
+ "reinforce_with_p_normalization": "MaxRL",
151
+ }
152
+
153
+ with open(SNR_STATS_PATH) as f:
154
+ plot_data = json.load(f)
155
+
156
+ fig, ax = plt.subplots(figsize=(8, 5))
157
+
158
+ for advantage_type, rollout_snrs in plot_data.items():
159
+ xs = [int(k) for k in rollout_snrs]
160
+ ys = [v["snr"] if isinstance(v, dict) else v for v in rollout_snrs.values()]
161
+ label = LABEL_MAP.get(advantage_type, advantage_type)
162
+ ax.plot(xs, ys, marker="o", label=label)
163
+
164
+ ax.set_xscale("log", base=2)
165
+ ax.set_xlabel("Rollouts (N)")
166
+ ax.set_ylabel("Gradient SNR")
167
+ ax.set_title(f"Gradient SNR vs Rollouts (S={S} gradient samples)")
168
+ ax.legend()
169
+ ax.grid(True, which="both", linestyle="--", alpha=0.5)
170
+
171
+ plt.tight_layout()
172
+ plt.savefig("snr.png", dpi=150)
173
+ print("Saved plot to snr.png")
data/Maze/variance/0000.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2555071921d6b00d268ce92ecb9a83d95572f88da90bd6b43e7dd5c818681e1c
3
+ size 3839835920
data/Maze/variance/5000.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a220dcd2bb4f8deb217c261cd11591c4a50f19b34c5d2d4927b273bcb778dc7c
3
+ size 3874320653
data/Maze/variance/8000.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7454efc155e9dbfe987db40abfc89925d3c3e24eca383ad13b5ce83e58b12fcf
3
+ size 3943981305
data/Qwen3/1.7b_512x512.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5c57ae4437ad668abaab5f1f5b162e72d3b53a34fc25f9047223c9a4b5ee34c
3
+ size 1722861371
data/Qwen3/4b_1024x128.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06b435d223068ec13f85407068b385fd56cc8c29e001aaca3f42ff4e489a1222
3
+ size 1063867627
data/README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ This directory contains data of the following:
2
+
3
+ 1. SmolLM + GSM8k
4
+
5
+ - Model: [SmolLM_MaxRL_rollout128_step1000](https://huggingface.co/guanning-ai/smollm2_0.3B_MaxRL_gsm8k_1000_steps)
6
+ - Dataset: 512 prompts x 512 rollouts in GSM8k Training set
7
+
8
+ 2. Qwen Math
9
+
10
+ - Model: [Qwen3-1.7B_MaxRL_1000steps](https://huggingface.co/collections/ftajwar/maxrl)
11
+ - Dataset: 512 prompts x 512 rollouts in Polaris-53K dataset
12
+ - Model: [Qwen3-4B_MaxRL_1000steps](https://huggingface.co/collections/ftajwar/maxrl)
13
+ - Dataset: 1024 prompts x 128 rollouts in Polaris-53K dataset
14
+
15
+ 3. Maze
16
+
17
+ - Models: [Maze17_bz256_ns64(0,5000,8000steps)](https://huggingface.co/guanning-ai/Qwen2-3M_MaxRL_Maze17_bz256_ns64)
18
+ - Dataset: [Maze17_train_parquet](https://huggingface.co/datasets/guanning-ai/maze_17x17_1m), 2048 prompts x 1024 rollouts for each prompt
data/SmolLM/512x512.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42a837cf816ec490629bf0e81959ad430ed2bb6106c799efb2f7fd65c9d64067
3
+ size 294900706
maze_variance_analysis.py ADDED
File without changes
maze_variance_analysis_auto.py ADDED
@@ -0,0 +1,400 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Automated Maze Variance Analysis for MaxRL Policy Gradient.
3
+
4
+ Runs all 12 experiments (6 rollout_nums x 2 baseline settings) across multiple
5
+ GPUs with dynamic scheduling. Each GPU worker loads the model once and pulls
6
+ experiments from a shared task pool. Saves only trace_covariance mean/std to
7
+ outputs/ and plots the variance line chart.
8
+ """
9
+
10
+ import json
11
+ import os
12
+ import random
13
+ from functools import partial
14
+
15
+ import matplotlib
16
+ matplotlib.use("Agg")
17
+ import matplotlib.pyplot as plt
18
+ import numpy as np
19
+ import torch
20
+ import torch.multiprocessing as mp
21
+ from transformers import AutoModelForCausalLM, AutoTokenizer
22
+
23
+ # ============================================================================
24
+ # Configuration
25
+ # ============================================================================
26
+ BATCH_SIZE = 256
27
+ ROLLOUT_NUMS = [4, 8, 16, 32, 64, 128, 256]
28
+ NUMBER_BATCHES_PER_ROUND = 4
29
+ TOTAL_ROUNDS = 128
30
+ MICRO_BATCH_SIZE = 1024
31
+ MAX_SEQ_LEN = 512
32
+ SEED = 42
33
+
34
+ MODEL_PATH = "/work/nvme/bgif/gzeng/MAXRL/checkpoints/maze/Qwen2-3M_MaxRL_Maze17_bz256_ns64/step5000"
35
+ DATA_PATH = "/work/nvme/bgif/gzeng/MAXRL/variance_analysis/data/Maze/variance/5000.jsonl"
36
+
37
+ GPU_IDS = [0, 1, 2, 3]
38
+ DTYPE = torch.float32
39
+
40
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
41
+ OUTPUT_DIR = os.path.join(SCRIPT_DIR, "outputs", "Maze")
42
+
43
+
44
+ # ============================================================================
45
+ # Per-worker global state (initialized once per GPU worker)
46
+ # ============================================================================
47
+ _worker_model = None
48
+ _worker_tokenizer = None
49
+ _worker_prompt_data = None
50
+ _worker_all_prompt_ids = None
51
+ _worker_total_params = None
52
+ _worker_device = None
53
+
54
+
55
+ def worker_init(gpu_queue: mp.Queue):
56
+ """Called once per pool worker. Grabs a GPU and loads model + data."""
57
+ global _worker_model, _worker_tokenizer, _worker_prompt_data
58
+ global _worker_all_prompt_ids, _worker_total_params, _worker_device
59
+
60
+ gpu_id = gpu_queue.get()
61
+ _worker_device = f"cuda:{gpu_id}"
62
+ print(f"[Worker PID={os.getpid()}] Assigned to GPU {gpu_id}")
63
+
64
+ _worker_tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
65
+ if _worker_tokenizer.pad_token is None:
66
+ _worker_tokenizer.pad_token = _worker_tokenizer.eos_token
67
+
68
+ _worker_model = AutoModelForCausalLM.from_pretrained(
69
+ MODEL_PATH, torch_dtype=DTYPE,
70
+ ).to(_worker_device)
71
+ _worker_model.eval()
72
+ for p in _worker_model.parameters():
73
+ p.requires_grad_(True)
74
+
75
+ _worker_total_params = sum(p.numel() for p in _worker_model.parameters())
76
+ print(f"[GPU {gpu_id}] Model loaded: {_worker_total_params:,} parameters")
77
+
78
+ _worker_prompt_data = load_rollout_data(DATA_PATH)
79
+ _worker_all_prompt_ids = list(_worker_prompt_data.keys())
80
+
81
+
82
+ # ============================================================================
83
+ # Data Loading
84
+ # ============================================================================
85
+ def load_rollout_data(data_path: str) -> dict:
86
+ prompt_to_id = {}
87
+ prompt_data = {}
88
+
89
+ with open(data_path, "r") as f:
90
+ for line in f:
91
+ item = json.loads(line)
92
+ prompt_text = item["input"]
93
+ if prompt_text not in prompt_to_id:
94
+ pid = len(prompt_to_id)
95
+ prompt_to_id[prompt_text] = pid
96
+ prompt_data[pid] = {"input": prompt_text, "rollouts": []}
97
+ pid = prompt_to_id[prompt_text]
98
+ prompt_data[pid]["rollouts"].append({
99
+ "output": item["output"],
100
+ "score": item["score"],
101
+ })
102
+
103
+ print(f"Loaded {len(prompt_data)} prompts, "
104
+ f"each with {len(prompt_data[0]['rollouts'])} rollouts")
105
+ return prompt_data
106
+
107
+
108
+ # ============================================================================
109
+ # MaxRL Advantage Computation
110
+ # ============================================================================
111
+ def compute_maxrl_advantage(
112
+ scores: list[float], baseline: bool, epsilon: float = 1e-6,
113
+ ) -> list[float]:
114
+ mean = sum(scores) / len(scores)
115
+ if baseline:
116
+ return [(s - mean) / (mean + epsilon) for s in scores]
117
+ else:
118
+ return [s / (mean + epsilon) for s in scores]
119
+
120
+
121
+ # ============================================================================
122
+ # Tokenization & Batching
123
+ # ============================================================================
124
+ def tokenize_and_get_response_mask(
125
+ tokenizer, prompt: str, response: str, max_seq_len: int,
126
+ ) -> tuple[torch.Tensor, torch.Tensor]:
127
+ prompt_ids = tokenizer.encode(prompt, add_special_tokens=False)
128
+ response_ids = tokenizer.encode(response, add_special_tokens=False)
129
+
130
+ total_len = len(prompt_ids) + len(response_ids)
131
+ if total_len > max_seq_len:
132
+ max_resp = max_seq_len - len(prompt_ids)
133
+ if max_resp <= 0:
134
+ prompt_ids = prompt_ids[:max_seq_len // 2]
135
+ max_resp = max_seq_len - len(prompt_ids)
136
+ response_ids = response_ids[:max_resp]
137
+
138
+ input_ids = prompt_ids + response_ids
139
+ response_mask = [0] * len(prompt_ids) + [1] * len(response_ids)
140
+
141
+ return (
142
+ torch.tensor(input_ids, dtype=torch.long),
143
+ torch.tensor(response_mask, dtype=torch.float32),
144
+ )
145
+
146
+
147
+ def pad_batch(
148
+ batch_input_ids: list[torch.Tensor],
149
+ batch_response_masks: list[torch.Tensor],
150
+ pad_token_id: int,
151
+ ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
152
+ max_len = max(ids.shape[0] for ids in batch_input_ids)
153
+ B = len(batch_input_ids)
154
+
155
+ input_ids = torch.full((B, max_len), pad_token_id, dtype=torch.long)
156
+ response_mask = torch.zeros(B, max_len)
157
+ attention_mask = torch.zeros(B, max_len)
158
+
159
+ for i, (ids, rmask) in enumerate(zip(batch_input_ids, batch_response_masks)):
160
+ seq_len = ids.shape[0]
161
+ input_ids[i, max_len - seq_len:] = ids
162
+ response_mask[i, max_len - seq_len:] = rmask
163
+ attention_mask[i, max_len - seq_len:] = 1.0
164
+
165
+ return input_ids, response_mask, attention_mask
166
+
167
+
168
+ # ============================================================================
169
+ # Policy Gradient Loss
170
+ # ============================================================================
171
+ def compute_policy_gradient_loss(
172
+ model, input_ids, attention_mask, response_mask, advantages,
173
+ ) -> tuple[torch.Tensor, int]:
174
+ outputs = model(input_ids=input_ids, attention_mask=attention_mask)
175
+ logits = outputs.logits
176
+
177
+ shift_logits = logits[:, :-1, :]
178
+ shift_labels = input_ids[:, 1:]
179
+ shift_response_mask = response_mask[:, 1:]
180
+
181
+ log_probs = torch.log_softmax(shift_logits, dim=-1)
182
+ token_log_probs = torch.gather(
183
+ log_probs, dim=-1, index=shift_labels.unsqueeze(-1),
184
+ ).squeeze(-1)
185
+
186
+ token_losses = -advantages.unsqueeze(-1) * token_log_probs * shift_response_mask
187
+ valid_token_count = int(shift_response_mask.sum().item())
188
+ loss = token_losses.sum() / max(valid_token_count, 1)
189
+ return loss, valid_token_count
190
+
191
+
192
+ # ============================================================================
193
+ # Gradient Utilities
194
+ # ============================================================================
195
+ def collect_flat_gradient(model) -> torch.Tensor:
196
+ grads = []
197
+ for p in model.parameters():
198
+ if p.grad is not None:
199
+ grads.append(p.grad.detach().float().flatten())
200
+ else:
201
+ grads.append(torch.zeros(p.numel(), dtype=torch.float32, device=p.device))
202
+ return torch.cat(grads)
203
+
204
+
205
+ def compute_trace_variance(
206
+ grad_sum: torch.Tensor, grad_sq_sum: torch.Tensor, K: int,
207
+ ) -> float:
208
+ grad_mean = grad_sum / K
209
+ elementwise_var = (grad_sq_sum / K - grad_mean ** 2) * (K / (K - 1))
210
+ return elementwise_var.sum().item()
211
+
212
+
213
+ # ============================================================================
214
+ # Single Experiment (runs inside a worker process)
215
+ # ============================================================================
216
+ def run_single_experiment(task: tuple[int, bool]) -> tuple[str, dict]:
217
+ """Run one experiment using the worker's pre-loaded model and data.
218
+
219
+ Args:
220
+ task: (rollout_num, baseline)
221
+
222
+ Returns:
223
+ (key, {"mean": float, "std": float})
224
+ """
225
+ rollout_num, baseline = task
226
+ key = f"nr{rollout_num}_bl{baseline}"
227
+
228
+ model = _worker_model
229
+ tokenizer = _worker_tokenizer
230
+ prompt_data = _worker_prompt_data
231
+ all_prompt_ids = _worker_all_prompt_ids
232
+ total_params = _worker_total_params
233
+ device = _worker_device
234
+
235
+ print(f"[{device}] Starting {key}")
236
+
237
+ random.seed(SEED)
238
+ np.random.seed(SEED)
239
+ torch.manual_seed(SEED)
240
+
241
+ trace_variances = []
242
+
243
+ for round_idx in range(TOTAL_ROUNDS):
244
+ sampled_prompts = random.sample(all_prompt_ids, BATCH_SIZE)
245
+
246
+ rollouts_needed = NUMBER_BATCHES_PER_ROUND * rollout_num
247
+ round_rollout_subsets = {}
248
+ for pid in sampled_prompts:
249
+ rollouts = prompt_data[pid]["rollouts"]
250
+ if len(rollouts) < rollouts_needed:
251
+ raise ValueError(
252
+ f"Prompt {pid} has {len(rollouts)} rollouts, need {rollouts_needed}"
253
+ )
254
+ sampled = random.sample(rollouts, rollouts_needed)
255
+ round_rollout_subsets[pid] = [
256
+ sampled[s:s + rollout_num]
257
+ for s in range(0, rollouts_needed, rollout_num)
258
+ ]
259
+
260
+ grad_sum = torch.zeros(total_params, dtype=torch.float32)
261
+ grad_sq_sum = torch.zeros(total_params, dtype=torch.float32)
262
+
263
+ for subset_idx in range(NUMBER_BATCHES_PER_ROUND):
264
+ all_input_ids = []
265
+ all_response_masks = []
266
+ all_advantages = []
267
+
268
+ for pid in sampled_prompts:
269
+ prompt_text = prompt_data[pid]["input"]
270
+ sampled_rollouts = round_rollout_subsets[pid][subset_idx]
271
+ scores = [r["score"] for r in sampled_rollouts]
272
+ advantages = compute_maxrl_advantage(scores, baseline)
273
+
274
+ for rollout, adv in zip(sampled_rollouts, advantages):
275
+ ids, rmask = tokenize_and_get_response_mask(
276
+ tokenizer, prompt_text, rollout["output"], MAX_SEQ_LEN,
277
+ )
278
+ all_input_ids.append(ids)
279
+ all_response_masks.append(rmask)
280
+ all_advantages.append(adv)
281
+
282
+ model.zero_grad()
283
+ total_valid_tokens = int(
284
+ sum(rmask[1:].sum().item() for rmask in all_response_masks)
285
+ )
286
+ num_samples = len(all_input_ids)
287
+
288
+ for mb_start in range(0, num_samples, MICRO_BATCH_SIZE):
289
+ mb_end = min(mb_start + MICRO_BATCH_SIZE, num_samples)
290
+
291
+ mb_ids = all_input_ids[mb_start:mb_end]
292
+ mb_masks = all_response_masks[mb_start:mb_end]
293
+ mb_advs = all_advantages[mb_start:mb_end]
294
+
295
+ input_ids, response_mask, attention_mask = pad_batch(
296
+ mb_ids, mb_masks, tokenizer.pad_token_id,
297
+ )
298
+ input_ids = input_ids.to(device)
299
+ response_mask = response_mask.to(device)
300
+ attention_mask = attention_mask.to(device)
301
+ advantages_t = torch.tensor(mb_advs, dtype=DTYPE, device=device)
302
+
303
+ mb_loss, mb_valid_tokens = compute_policy_gradient_loss(
304
+ model, input_ids, attention_mask, response_mask, advantages_t,
305
+ )
306
+ scaled_loss = mb_loss * (mb_valid_tokens / max(total_valid_tokens, 1))
307
+ scaled_loss.backward()
308
+
309
+ flat_grad = collect_flat_gradient(model).cpu()
310
+ grad_sum += flat_grad
311
+ grad_sq_sum += flat_grad ** 2
312
+
313
+ trace_var = compute_trace_variance(
314
+ grad_sum, grad_sq_sum, NUMBER_BATCHES_PER_ROUND,
315
+ )
316
+ trace_variances.append(trace_var)
317
+ print(f" [{device}] {key} round {round_idx+1}/{TOTAL_ROUNDS}: "
318
+ f"trace_cov={trace_var:.6e}")
319
+
320
+ result = {
321
+ "mean": float(np.mean(trace_variances)),
322
+ "std": float(np.std(trace_variances)),
323
+ }
324
+ print(f"[{device}] Finished {key}: mean={result['mean']:.6e}, std={result['std']:.6e}")
325
+ return key, result
326
+
327
+
328
+ # ============================================================================
329
+ # Plotting
330
+ # ============================================================================
331
+ def plot_results(results: dict, output_dir: str):
332
+ rollout_nums = ROLLOUT_NUMS
333
+
334
+ means_bl_true = [results[f"nr{nr}_blTrue"]["mean"] for nr in rollout_nums]
335
+ means_bl_false = [results[f"nr{nr}_blFalse"]["mean"] for nr in rollout_nums]
336
+
337
+ fig, ax = plt.subplots(figsize=(7, 5))
338
+
339
+ ax.plot(rollout_nums, means_bl_true, marker='o', label='MaxRL')
340
+ ax.plot(rollout_nums, means_bl_false, marker='s', label='MaxRL (w/o baseline)')
341
+
342
+ ax.set_xscale('log', base=2)
343
+ ax.set_xticks(rollout_nums)
344
+ ax.set_xticklabels(rollout_nums)
345
+ ax.set_xlabel('Rollout', fontsize=14)
346
+ ax.set_ylabel('Gradient Variance', fontsize=14)
347
+ ax.legend(fontsize=12)
348
+ ax.grid(True, alpha=0.3)
349
+
350
+ plt.tight_layout()
351
+ plt.savefig(os.path.join(output_dir, "variance_plot.pdf"), dpi=300)
352
+ plt.savefig(os.path.join(output_dir, "variance_plot.png"), dpi=300)
353
+ print(f"Plots saved to {output_dir}/variance_plot.{{pdf,png}}")
354
+
355
+
356
+ # ============================================================================
357
+ # Main
358
+ # ============================================================================
359
+ def main():
360
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
361
+
362
+ # Build task list: 12 experiments
363
+ tasks = []
364
+ for rollout_num in ROLLOUT_NUMS:
365
+ for baseline in [True, False]:
366
+ tasks.append((rollout_num, baseline))
367
+
368
+ print(f"Scheduling {len(tasks)} experiments across {len(GPU_IDS)} GPUs")
369
+
370
+ # GPU queue: each worker grabs one GPU ID on init
371
+ gpu_queue = mp.Queue()
372
+ for gid in GPU_IDS:
373
+ gpu_queue.put(gid)
374
+
375
+ # Pool of workers = number of GPUs. Each worker inits once (loads model),
376
+ # then processes tasks dynamically from the pool.
377
+ with mp.Pool(
378
+ processes=len(GPU_IDS),
379
+ initializer=worker_init,
380
+ initargs=(gpu_queue,),
381
+ ) as pool:
382
+ results_list = pool.map(run_single_experiment, tasks)
383
+
384
+ # Collect results
385
+ results = dict(results_list)
386
+
387
+ # Save
388
+ results_path = os.path.join(OUTPUT_DIR, "results.json")
389
+ with open(results_path, "w") as f:
390
+ json.dump(results, f, indent=2)
391
+ print(f"Results saved to {results_path}")
392
+
393
+ # Plot
394
+ plot_results(results, OUTPUT_DIR)
395
+ print("All done!")
396
+
397
+
398
+ if __name__ == "__main__":
399
+ mp.set_start_method("spawn")
400
+ main()
outputs/Maze/results.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "nr4_blTrue": {
3
+ "mean": 0.09343412215821445,
4
+ "std": 0.04141906356644995
5
+ },
6
+ "nr4_blFalse": {
7
+ "mean": 0.1171639968524687,
8
+ "std": 0.06142266485721379
9
+ },
10
+ "nr8_blTrue": {
11
+ "mean": 0.07531368439958896,
12
+ "std": 0.037395865364248826
13
+ },
14
+ "nr8_blFalse": {
15
+ "mean": 0.0812347744795261,
16
+ "std": 0.05446751649965604
17
+ },
18
+ "nr16_blTrue": {
19
+ "mean": 0.051791880556265824,
20
+ "std": 0.032323674211987063
21
+ },
22
+ "nr16_blFalse": {
23
+ "mean": 0.056146801136492286,
24
+ "std": 0.03758377689938391
25
+ },
26
+ "nr32_blTrue": {
27
+ "mean": 0.04027390201736125,
28
+ "std": 0.03491190177211898
29
+ },
30
+ "nr32_blFalse": {
31
+ "mean": 0.04236685928481165,
32
+ "std": 0.03529769154208022
33
+ },
34
+ "nr64_blTrue": {
35
+ "mean": 0.02930887994079967,
36
+ "std": 0.029723109116621262
37
+ },
38
+ "nr64_blFalse": {
39
+ "mean": 0.030155911130350432,
40
+ "std": 0.02806844767402451
41
+ },
42
+ "nr128_blTrue": {
43
+ "mean": 0.023548210993794783,
44
+ "std": 0.027154330195472722
45
+ },
46
+ "nr128_blFalse": {
47
+ "mean": 0.02521930770672043,
48
+ "std": 0.03021973835553808
49
+ },
50
+ "nr256_blTrue": {
51
+ "mean": 0.017309359611317632,
52
+ "std": 0.01862850514229902
53
+ },
54
+ "nr256_blFalse": {
55
+ "mean": 0.017515211292220556,
56
+ "std": 0.018245087061411355
57
+ }
58
+ }
outputs/Maze/variance_plot.pdf ADDED
Binary file (14.5 kB). View file
 
outputs/Maze/variance_plot.png ADDED

Git LFS Details

  • SHA256: 9d83e91dd1b198748c1923042f1a14035ccc06aa825ca2bd4e1f74a47ffa507a
  • Pointer size: 131 Bytes
  • Size of remote file: 149 kB
outputs/Qwen3/.gitkeep ADDED
@@ -0,0 +1 @@
 
 
1
+
outputs/SmolLM/results.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "nr4_blTrue": {
3
+ "mean": 0.3461569035425782,
4
+ "std": 0.10700828370370079
5
+ },
6
+ "nr4_blFalse": {
7
+ "mean": 0.5613684519194067,
8
+ "std": 0.1363638089093747
9
+ },
10
+ "nr8_blTrue": {
11
+ "mean": 0.34889423102140427,
12
+ "std": 0.11797077527297667
13
+ },
14
+ "nr8_blFalse": {
15
+ "mean": 0.46802500914782286,
16
+ "std": 0.13111071007542788
17
+ },
18
+ "nr16_blTrue": {
19
+ "mean": 0.27634103666059673,
20
+ "std": 0.09890688698321183
21
+ },
22
+ "nr16_blFalse": {
23
+ "mean": 0.345704042352736,
24
+ "std": 0.10513549380879457
25
+ },
26
+ "nr32_blTrue": {
27
+ "mean": 0.26002368051558733,
28
+ "std": 0.12564290483305135
29
+ },
30
+ "nr32_blFalse": {
31
+ "mean": 0.29659369541332126,
32
+ "std": 0.13365507973097127
33
+ },
34
+ "nr64_blTrue": {
35
+ "mean": 0.2039303722558543,
36
+ "std": 0.1140506321981112
37
+ },
38
+ "nr64_blFalse": {
39
+ "mean": 0.2275285553187132,
40
+ "std": 0.11667669231022332
41
+ },
42
+ "nr128_blTrue": {
43
+ "mean": 0.1596078732982278,
44
+ "std": 0.07754145558626176
45
+ },
46
+ "nr128_blFalse": {
47
+ "mean": 0.17232763615902513,
48
+ "std": 0.07963091239252962
49
+ }
50
+ }
outputs/SmolLM/variance_plot.pdf ADDED
Binary file (14.5 kB). View file
 
outputs/SmolLM/variance_plot.png ADDED

Git LFS Details

  • SHA256: b82d33fe16218534a4ee1b4382fea6a22c5d796aec657ff1c42c466b16b5f605
  • Pointer size: 131 Bytes
  • Size of remote file: 149 kB
outputs/SmolLM_SNR/.gitkeep ADDED
@@ -0,0 +1 @@
 
 
1
+
outputs/old/results_bs16_nr128_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 128,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.7689405679702759,
15
+ "trace_variance": 0.2869313657283783,
16
+ "relative_variance": 0.4852800940577739,
17
+ "avg_sample_grad_norm": 0.8799326419830322,
18
+ "std_sample_grad_norm": 0.1096991376772517,
19
+ "avg_cosine_similarity_to_mean": 0.9405284523963928
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.8357224464416504,
23
+ "trace_variance": 0.15966498851776123,
24
+ "relative_variance": 0.22860491326781845,
25
+ "avg_sample_grad_norm": 0.8944469094276428,
26
+ "std_sample_grad_norm": 0.07733634177736912,
27
+ "avg_cosine_similarity_to_mean": 0.9950093924999237
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.6839278936386108,
31
+ "trace_variance": 0.08836942911148071,
32
+ "relative_variance": 0.18892151348330627,
33
+ "avg_sample_grad_norm": 0.725984051823616,
34
+ "std_sample_grad_norm": 0.029560648578731755,
35
+ "avg_cosine_similarity_to_mean": 1.011790782213211
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.7902998328208923,
39
+ "trace_variance": 0.06472273916006088,
40
+ "relative_variance": 0.1036270437226223,
41
+ "avg_sample_grad_norm": 0.8173244595527649,
42
+ "std_sample_grad_norm": 0.020897730886255742,
43
+ "avg_cosine_similarity_to_mean": 1.0311751067638397
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.8064079284667969,
47
+ "trace_variance": 0.15503337979316711,
48
+ "relative_variance": 0.2384051522286745,
49
+ "avg_sample_grad_norm": 0.868552565574646,
50
+ "std_sample_grad_norm": 0.03490493615590108,
51
+ "avg_cosine_similarity_to_mean": 0.9947420358657837
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.7770597338676453,
57
+ "std": 0.05141476657999565
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.15094438046216965,
61
+ "std": 0.07737573915323949
62
+ },
63
+ "relative_variance": {
64
+ "mean": 0.2489677433520391,
65
+ "std": 0.12735714268667106
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.8372481256723404,
69
+ "std": 0.06140078721713483
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.05447975901510188,
73
+ "std": 0.03377422544946952
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.9946491539478302,
77
+ "std": 0.03018996317980766
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr128_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 128,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.3161277174949646,
15
+ "trace_variance": 0.27833467721939087,
16
+ "relative_variance": 2.7851088027893787,
17
+ "avg_sample_grad_norm": 0.5257724076509476,
18
+ "std_sample_grad_norm": 0.1253468533977172,
19
+ "avg_cosine_similarity_to_mean": 0.6508230715990067
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.23199908435344696,
23
+ "trace_variance": 0.15211263298988342,
24
+ "relative_variance": 2.8261339442659383,
25
+ "avg_sample_grad_norm": 0.3839375451207161,
26
+ "std_sample_grad_norm": 0.10596379740421964,
27
+ "avg_cosine_similarity_to_mean": 0.6401898264884949
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.22157494723796844,
31
+ "trace_variance": 0.07338934391736984,
32
+ "relative_variance": 1.4948296214837793,
33
+ "avg_sample_grad_norm": 0.3141995221376419,
34
+ "std_sample_grad_norm": 0.02690045819675988,
35
+ "avg_cosine_similarity_to_mean": 0.7678444683551788
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.2028011679649353,
39
+ "trace_variance": 0.05692358687520027,
40
+ "relative_variance": 1.3840486413650959,
41
+ "avg_sample_grad_norm": 0.2825518846511841,
42
+ "std_sample_grad_norm": 0.01270163227318277,
43
+ "avg_cosine_similarity_to_mean": 0.7714781761169434
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.2694803774356842,
47
+ "trace_variance": 0.14570380747318268,
48
+ "relative_variance": 2.0063957823319107,
49
+ "avg_sample_grad_norm": 0.4091017544269562,
50
+ "std_sample_grad_norm": 0.07354999374759726,
51
+ "avg_cosine_similarity_to_mean": 0.715828612446785
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.2483966588973999,
57
+ "std": 0.040247367483568344
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.1412928096950054,
61
+ "std": 0.07828926536369483
62
+ },
63
+ "relative_variance": {
64
+ "mean": 2.0993033584472203,
65
+ "std": 0.613876442723994
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.38311262279748914,
69
+ "std": 0.08472237859410245
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.06889254700389535,
73
+ "std": 0.043597735141821393
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.7092328310012818,
77
+ "std": 0.05573499407075196
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr16_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 16,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.6964784264564514,
15
+ "trace_variance": 0.3348350524902344,
16
+ "relative_variance": 0.6902645643219107,
17
+ "avg_sample_grad_norm": 0.8420403301715851,
18
+ "std_sample_grad_norm": 0.07201185501512579,
19
+ "avg_cosine_similarity_to_mean": 0.8888210356235504
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.8008900880813599,
23
+ "trace_variance": 0.4339163601398468,
24
+ "relative_variance": 0.6764881401105687,
25
+ "avg_sample_grad_norm": 0.9527603536844254,
26
+ "std_sample_grad_norm": 0.1757878804217989,
27
+ "avg_cosine_similarity_to_mean": 0.8986288011074066
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.7970862984657288,
31
+ "trace_variance": 0.4751781225204468,
32
+ "relative_variance": 0.7479038165554954,
33
+ "avg_sample_grad_norm": 0.9778163135051727,
34
+ "std_sample_grad_norm": 0.05700159387678907,
35
+ "avg_cosine_similarity_to_mean": 0.8750215619802475
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.6043644547462463,
39
+ "trace_variance": 0.35235679149627686,
40
+ "relative_variance": 0.9646834306882995,
41
+ "avg_sample_grad_norm": 0.7641780376434326,
42
+ "std_sample_grad_norm": 0.15278308552449663,
43
+ "avg_cosine_similarity_to_mean": 0.8460912108421326
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.5035268068313599,
47
+ "trace_variance": 0.30590248107910156,
48
+ "relative_variance": 1.2065291143381554,
49
+ "avg_sample_grad_norm": 0.6727220565080643,
50
+ "std_sample_grad_norm": 0.09560943366609909,
51
+ "avg_cosine_similarity_to_mean": 0.8143545389175415
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.6804692149162292,
57
+ "std": 0.11441815941797533
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.3804377615451813,
61
+ "std": 0.06365430463302185
62
+ },
63
+ "relative_variance": {
64
+ "mean": 0.8571738132028859,
65
+ "std": 0.20300413112709334
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.841903418302536,
69
+ "std": 0.11438982640665354
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.11063876970086188,
73
+ "std": 0.04607692718517292
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.8645834296941757,
77
+ "std": 0.030725907922132888
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr16_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 16,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.2737724781036377,
15
+ "trace_variance": 0.2792177200317383,
16
+ "relative_variance": 3.7253184368368504,
17
+ "avg_sample_grad_norm": 0.5060313940048218,
18
+ "std_sample_grad_norm": 0.10218939775503928,
19
+ "avg_cosine_similarity_to_mean": 0.5696069896221161
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.32015320658683777,
23
+ "trace_variance": 0.31460481882095337,
24
+ "relative_variance": 3.069372926299306,
25
+ "avg_sample_grad_norm": 0.541322261095047,
26
+ "std_sample_grad_norm": 0.163643840876535,
27
+ "avg_cosine_similarity_to_mean": 0.5857272818684578
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.377981960773468,
31
+ "trace_variance": 0.41017645597457886,
32
+ "relative_variance": 2.870969514051388,
33
+ "avg_sample_grad_norm": 0.6466762721538544,
34
+ "std_sample_grad_norm": 0.08176618746444127,
35
+ "avg_cosine_similarity_to_mean": 0.6339955478906631
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.2758277654647827,
39
+ "trace_variance": 0.2821664810180664,
40
+ "relative_variance": 3.708766223976379,
41
+ "avg_sample_grad_norm": 0.5054576545953751,
42
+ "std_sample_grad_norm": 0.12428782340874198,
43
+ "avg_cosine_similarity_to_mean": 0.5777123421430588
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.2792853116989136,
47
+ "trace_variance": 0.2728402018547058,
48
+ "relative_variance": 3.497938505582848,
49
+ "avg_sample_grad_norm": 0.5002564936876297,
50
+ "std_sample_grad_norm": 0.11702612906514184,
51
+ "avg_cosine_similarity_to_mean": 0.5859523713588715
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.30540414452552794,
57
+ "std": 0.04010633912538859
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.31180113554000854,
61
+ "std": 0.051270674618136725
62
+ },
63
+ "relative_variance": {
64
+ "mean": 3.3744731213493546,
65
+ "std": 0.34545333304198694
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.5399488151073456,
69
+ "std": 0.055331995136430764
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.11778267571397986,
73
+ "std": 0.027153171169363003
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.5905989065766335,
77
+ "std": 0.022518062147937085
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr32_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 32,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.7721027135848999,
15
+ "trace_variance": 0.3138512372970581,
16
+ "relative_variance": 0.526470071209843,
17
+ "avg_sample_grad_norm": 0.8968465626239777,
18
+ "std_sample_grad_norm": 0.08699080250453943,
19
+ "avg_cosine_similarity_to_mean": 0.9272608906030655
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.76179438829422,
23
+ "trace_variance": 0.2826382517814636,
24
+ "relative_variance": 0.48702964807005555,
25
+ "avg_sample_grad_norm": 0.8696170151233673,
26
+ "std_sample_grad_norm": 0.12881966504374284,
27
+ "avg_cosine_similarity_to_mean": 0.9404138624668121
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.6628957390785217,
31
+ "trace_variance": 0.35148754715919495,
32
+ "relative_variance": 0.7998701464789485,
33
+ "avg_sample_grad_norm": 0.8236571550369263,
34
+ "std_sample_grad_norm": 0.024779726969064976,
35
+ "avg_cosine_similarity_to_mean": 0.8674887269735336
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.8706740736961365,
39
+ "trace_variance": 0.3642888367176056,
40
+ "relative_variance": 0.48054563614774876,
41
+ "avg_sample_grad_norm": 1.0016238987445831,
42
+ "std_sample_grad_norm": 0.03215233180573083,
43
+ "avg_cosine_similarity_to_mean": 0.9396774470806122
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.6275550723075867,
47
+ "trace_variance": 0.30897197127342224,
48
+ "relative_variance": 0.784540549413837,
49
+ "avg_sample_grad_norm": 0.7746196091175079,
50
+ "std_sample_grad_norm": 0.05964617988155661,
51
+ "avg_cosine_similarity_to_mean": 0.8722260743379593
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.739004397392273,
57
+ "std": 0.08621515622002149
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.3242475688457489,
61
+ "std": 0.02972568723754252
62
+ },
63
+ "relative_variance": {
64
+ "mean": 0.6156912102640866,
65
+ "std": 0.14505896409400237
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.8732728481292724,
69
+ "std": 0.0764686287213851
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.06647774124092694,
73
+ "std": 0.0381337894879539
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.9094134002923966,
77
+ "std": 0.032668178822674025
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr32_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 32,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.28896379470825195,
15
+ "trace_variance": 0.2879572808742523,
16
+ "relative_variance": 3.448587102812029,
17
+ "avg_sample_grad_norm": 0.505200058221817,
18
+ "std_sample_grad_norm": 0.16691583237670862,
19
+ "avg_cosine_similarity_to_mean": 0.5931130349636078
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.2619099020957947,
23
+ "trace_variance": 0.23625493049621582,
24
+ "relative_variance": 3.444110237189513,
25
+ "avg_sample_grad_norm": 0.45509156957268715,
26
+ "std_sample_grad_norm": 0.15576454242091092,
27
+ "avg_cosine_similarity_to_mean": 0.5868209525942802
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.30585452914237976,
31
+ "trace_variance": 0.3292638063430786,
32
+ "relative_variance": 3.519768999971907,
33
+ "avg_sample_grad_norm": 0.5639973282814026,
34
+ "std_sample_grad_norm": 0.01767149544944113,
35
+ "avg_cosine_similarity_to_mean": 0.592338502407074
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.3031623363494873,
39
+ "trace_variance": 0.3239123523235321,
40
+ "relative_variance": 3.5243336693775595,
41
+ "avg_sample_grad_norm": 0.554802268743515,
42
+ "std_sample_grad_norm": 0.07338316736266386,
43
+ "avg_cosine_similarity_to_mean": 0.6006623655557632
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.2689555287361145,
47
+ "trace_variance": 0.27706390619277954,
48
+ "relative_variance": 3.830178373994615,
49
+ "avg_sample_grad_norm": 0.5079697072505951,
50
+ "std_sample_grad_norm": 0.05607249553495662,
51
+ "avg_cosine_similarity_to_mean": 0.5773692056536674
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.28576921820640566,
57
+ "std": 0.017709391361814476
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.2908904552459717,
61
+ "std": 0.03390509488651892
62
+ },
63
+ "relative_variance": {
64
+ "mean": 3.553395676669125,
65
+ "std": 0.14248660261828003
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.5174121864140033,
69
+ "std": 0.03921823420458731
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.09396150662893624,
73
+ "std": 0.058001999780733754
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.5900608122348785,
77
+ "std": 0.007726194885257127
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr4_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 4,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.7435137629508972,
15
+ "trace_variance": 0.6554153561592102,
16
+ "relative_variance": 1.1856010859189596,
17
+ "avg_sample_grad_norm": 0.996780276298523,
18
+ "std_sample_grad_norm": 0.09697026173657458,
19
+ "avg_cosine_similarity_to_mean": 0.8014578223228455
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.6562896370887756,
23
+ "trace_variance": 0.6258231401443481,
24
+ "relative_variance": 1.4529829690553044,
25
+ "avg_sample_grad_norm": 0.9213704764842987,
26
+ "std_sample_grad_norm": 0.11367962565706029,
27
+ "avg_cosine_similarity_to_mean": 0.767962858080864
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.6826515197753906,
31
+ "trace_variance": 0.5123142004013062,
32
+ "relative_variance": 1.0993557972789034,
33
+ "avg_sample_grad_norm": 0.8889862895011902,
34
+ "std_sample_grad_norm": 0.1649224690024605,
35
+ "avg_cosine_similarity_to_mean": 0.81678107380867
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.5915200114250183,
39
+ "trace_variance": 0.6056228876113892,
40
+ "relative_variance": 1.730865797634316,
41
+ "avg_sample_grad_norm": 0.8678804934024811,
42
+ "std_sample_grad_norm": 0.11112772837588089,
43
+ "avg_cosine_similarity_to_mean": 0.73161780834198
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.780635416507721,
47
+ "trace_variance": 0.5549829006195068,
48
+ "relative_variance": 0.9107162812869151,
49
+ "avg_sample_grad_norm": 0.9906752854585648,
50
+ "std_sample_grad_norm": 0.08322465784960068,
51
+ "avg_cosine_similarity_to_mean": 0.8424067795276642
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.6909220695495606,
57
+ "std": 0.06627661332380856
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.5908316969871521,
61
+ "std": 0.05111626535159937
62
+ },
63
+ "relative_variance": {
64
+ "mean": 1.2759043862348798,
65
+ "std": 0.28671698682220986
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.9331385642290115,
69
+ "std": 0.05235889392604594
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.11398494852431537,
73
+ "std": 0.02770769918405064
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.7920452684164048,
77
+ "std": 0.03864758326071824
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr4_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 4,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.3311220407485962,
15
+ "trace_variance": 0.4140247106552124,
16
+ "relative_variance": 3.7761573424860275,
17
+ "avg_sample_grad_norm": 0.6195628046989441,
18
+ "std_sample_grad_norm": 0.10239149261311377,
19
+ "avg_cosine_similarity_to_mean": 0.5686733275651932
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.509647786617279,
23
+ "trace_variance": 0.6342702507972717,
24
+ "relative_variance": 2.4419347611122593,
25
+ "avg_sample_grad_norm": 0.7901830524206161,
26
+ "std_sample_grad_norm": 0.2769427565102714,
27
+ "avg_cosine_similarity_to_mean": 0.6417316198348999
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.2220628559589386,
31
+ "trace_variance": 0.24883511662483215,
32
+ "relative_variance": 5.046146175333274,
33
+ "avg_sample_grad_norm": 0.4583381339907646,
34
+ "std_sample_grad_norm": 0.09868455803698328,
35
+ "avg_cosine_similarity_to_mean": 0.5111349001526833
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.3061383068561554,
39
+ "trace_variance": 0.38140782713890076,
40
+ "relative_variance": 4.069623654264621,
41
+ "avg_sample_grad_norm": 0.5863617360591888,
42
+ "std_sample_grad_norm": 0.10806355451567583,
43
+ "avg_cosine_similarity_to_mean": 0.5520333349704742
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.23846779763698578,
47
+ "trace_variance": 0.2480362355709076,
48
+ "relative_variance": 4.361698572068804,
49
+ "avg_sample_grad_norm": 0.47311121225357056,
50
+ "std_sample_grad_norm": 0.06102951740547904,
51
+ "avg_cosine_similarity_to_mean": 0.5363884419202805
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.321487757563591,
57
+ "std": 0.10247950662284001
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.3853148281574249,
61
+ "std": 0.14162658515570797
62
+ },
63
+ "relative_variance": {
64
+ "mean": 3.9391121010529973,
65
+ "std": 0.8589797754541765
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.5855113878846169,
69
+ "std": 0.11988123150089669
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.12942237581630464,
73
+ "std": 0.07559302499255817
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.5619923248887062,
77
+ "std": 0.044145688145756264
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr64_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 64,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.7199326157569885,
15
+ "trace_variance": 0.16369977593421936,
16
+ "relative_variance": 0.3158380040035924,
17
+ "avg_sample_grad_norm": 0.7896079123020172,
18
+ "std_sample_grad_norm": 0.08840608429535818,
19
+ "avg_cosine_similarity_to_mean": 0.9801101982593536
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.8231773972511292,
23
+ "trace_variance": 0.6244240999221802,
24
+ "relative_variance": 0.9214945738571112,
25
+ "avg_sample_grad_norm": 1.0009050518274307,
26
+ "std_sample_grad_norm": 0.3319705070724497,
27
+ "avg_cosine_similarity_to_mean": 0.8798657655715942
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.9061974883079529,
31
+ "trace_variance": 0.18488027155399323,
32
+ "relative_variance": 0.22513595665361935,
33
+ "avg_sample_grad_norm": 0.9721819162368774,
34
+ "std_sample_grad_norm": 0.04711759110632938,
35
+ "avg_cosine_similarity_to_mean": 1.003186285495758
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.8444343209266663,
39
+ "trace_variance": 0.1493278592824936,
40
+ "relative_variance": 0.2094156270354833,
41
+ "avg_sample_grad_norm": 0.9021681249141693,
42
+ "std_sample_grad_norm": 0.023769549794762847,
43
+ "avg_cosine_similarity_to_mean": 1.0020580887794495
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.5821331143379211,
47
+ "trace_variance": 0.21182984113693237,
48
+ "relative_variance": 0.6250899711173901,
49
+ "avg_sample_grad_norm": 0.6914167404174805,
50
+ "std_sample_grad_norm": 0.07562904595760037,
51
+ "avg_cosine_similarity_to_mean": 0.9062196463346481
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.7751749873161315,
57
+ "std": 0.11365286336587271
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.26683236956596373,
61
+ "std": 0.1800316149610794
62
+ },
63
+ "relative_variance": {
64
+ "mean": 0.4593948265334392,
65
+ "std": 0.27530580706159213
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.871255949139595,
69
+ "std": 0.11572298491365285
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.11337855564530011,
73
+ "std": 0.11158080514754416
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.9542879968881607,
77
+ "std": 0.05135960519860018
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr64_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 64,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.2755725681781769,
15
+ "trace_variance": 0.13866141438484192,
16
+ "relative_variance": 1.8259280401992952,
17
+ "avg_sample_grad_norm": 0.38493604958057404,
18
+ "std_sample_grad_norm": 0.15309434629388827,
19
+ "avg_cosine_similarity_to_mean": 0.7401671260595322
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.4243953227996826,
23
+ "trace_variance": 0.6096072793006897,
24
+ "relative_variance": 3.3846125939842477,
25
+ "avg_sample_grad_norm": 0.6765449717640877,
26
+ "std_sample_grad_norm": 0.3855063964765728,
27
+ "avg_cosine_similarity_to_mean": 0.5657118707895279
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.27996593713760376,
31
+ "trace_variance": 0.16036519408226013,
32
+ "relative_variance": 2.0459721790598753,
33
+ "avg_sample_grad_norm": 0.4325138255953789,
34
+ "std_sample_grad_norm": 0.040878211355412856,
35
+ "avg_cosine_similarity_to_mean": 0.7004560381174088
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.2685435116291046,
39
+ "trace_variance": 0.13540661334991455,
40
+ "relative_variance": 1.8776323020858432,
41
+ "avg_sample_grad_norm": 0.4076094403862953,
42
+ "std_sample_grad_norm": 0.01268766293214707,
43
+ "avg_cosine_similarity_to_mean": 0.7097402513027191
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.23571749031543732,
47
+ "trace_variance": 0.1872667372226715,
48
+ "relative_variance": 3.370365696985525,
49
+ "avg_sample_grad_norm": 0.4250478520989418,
50
+ "std_sample_grad_norm": 0.06120551716417047,
51
+ "avg_cosine_similarity_to_mean": 0.6070868670940399
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.29683896601200105,
57
+ "std": 0.06564152720226617
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.24626144766807556,
61
+ "std": 0.18261724839656943
62
+ },
63
+ "relative_variance": {
64
+ "mean": 2.5009021624629573,
65
+ "std": 0.719434171974512
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.4653304278850555,
69
+ "std": 0.10687017105313018
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.13067442684443828,
73
+ "std": 0.1358323611821015
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.6646324306726455,
77
+ "std": 0.06651361604169471
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr8_nb4_r5_blFalse.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 8,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": false
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.7105476260185242,
15
+ "trace_variance": 0.4447573125362396,
16
+ "relative_variance": 0.880920489967786,
17
+ "avg_sample_grad_norm": 0.8949873298406601,
18
+ "std_sample_grad_norm": 0.08978978326506477,
19
+ "avg_cosine_similarity_to_mean": 0.8567404597997665
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.7531298398971558,
23
+ "trace_variance": 0.5207900404930115,
24
+ "relative_variance": 0.91816970637431,
25
+ "avg_sample_grad_norm": 0.9467365890741348,
26
+ "std_sample_grad_norm": 0.1636269119236902,
27
+ "avg_cosine_similarity_to_mean": 0.8563707917928696
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.6841264963150024,
31
+ "trace_variance": 0.3036300837993622,
32
+ "relative_variance": 0.6487419431051044,
33
+ "avg_sample_grad_norm": 0.8194261938333511,
34
+ "std_sample_grad_norm": 0.06866591628645671,
35
+ "avg_cosine_similarity_to_mean": 0.9007236510515213
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.6350532174110413,
39
+ "trace_variance": 0.2895530164241791,
40
+ "relative_variance": 0.7179725695195097,
41
+ "avg_sample_grad_norm": 0.762009471654892,
42
+ "std_sample_grad_norm": 0.13810664175806275,
43
+ "avg_cosine_similarity_to_mean": 0.8990585058927536
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.47783538699150085,
47
+ "trace_variance": 0.38107824325561523,
48
+ "relative_variance": 1.669004609419634,
49
+ "avg_sample_grad_norm": 0.6930959820747375,
50
+ "std_sample_grad_norm": 0.09961858263046741,
51
+ "avg_cosine_similarity_to_mean": 0.7430207431316376
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.6521385133266449,
57
+ "std": 0.0951920067340915
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.38796173930168154,
61
+ "std": 0.08684766215895035
62
+ },
63
+ "relative_variance": {
64
+ "mean": 0.966961863677269,
65
+ "std": 0.3649403775708313
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.8232511132955551,
69
+ "std": 0.09066451825859152
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.11196156717274837,
73
+ "std": 0.034267545660677176
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.8511828303337097,
77
+ "std": 0.05745121848128472
78
+ }
79
+ }
80
+ }
outputs/old/results_bs16_nr8_nb4_r5_blTrue.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "batch_size": 16,
4
+ "rollout_num": 8,
5
+ "number_batches_per_round": 4,
6
+ "total_rounds": 5,
7
+ "model_path": "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps",
8
+ "max_seq_len": 2048,
9
+ "seed": 42,
10
+ "baseline": true
11
+ },
12
+ "per_round": [
13
+ {
14
+ "mean_grad_norm": 0.3068431317806244,
15
+ "trace_variance": 0.35510000586509705,
16
+ "relative_variance": 3.7715325967547115,
17
+ "avg_sample_grad_norm": 0.5735703259706497,
18
+ "std_sample_grad_norm": 0.09419738943035173,
19
+ "avg_cosine_similarity_to_mean": 0.5783642008900642
20
+ },
21
+ {
22
+ "mean_grad_norm": 0.2915055751800537,
23
+ "trace_variance": 0.35840773582458496,
24
+ "relative_variance": 4.217777287335315,
25
+ "avg_sample_grad_norm": 0.5577257871627808,
26
+ "std_sample_grad_norm": 0.13915088584464044,
27
+ "avg_cosine_similarity_to_mean": 0.5441212877631187
28
+ },
29
+ {
30
+ "mean_grad_norm": 0.21999073028564453,
31
+ "trace_variance": 0.2043260633945465,
32
+ "relative_variance": 4.2219686496831015,
33
+ "avg_sample_grad_norm": 0.4286266788840294,
34
+ "std_sample_grad_norm": 0.07468024966573314,
35
+ "avg_cosine_similarity_to_mean": 0.5293085053563118
36
+ },
37
+ {
38
+ "mean_grad_norm": 0.22899571061134338,
39
+ "trace_variance": 0.16267657279968262,
40
+ "relative_variance": 3.1022037496568386,
41
+ "avg_sample_grad_norm": 0.39560337364673615,
42
+ "std_sample_grad_norm": 0.08369558175357744,
43
+ "avg_cosine_similarity_to_mean": 0.6156094446778297
44
+ },
45
+ {
46
+ "mean_grad_norm": 0.29250001907348633,
47
+ "trace_variance": 0.3239949941635132,
48
+ "relative_variance": 3.7869232409130804,
49
+ "avg_sample_grad_norm": 0.5501251593232155,
50
+ "std_sample_grad_norm": 0.07720123505079711,
51
+ "avg_cosine_similarity_to_mean": 0.5774728059768677
52
+ }
53
+ ],
54
+ "averaged": {
55
+ "mean_grad_norm": {
56
+ "mean": 0.26796703338623046,
57
+ "std": 0.03602157970135398
58
+ },
59
+ "trace_variance": {
60
+ "mean": 0.28090107440948486,
61
+ "std": 0.08149922086187118
62
+ },
63
+ "relative_variance": {
64
+ "mean": 3.82008110486861,
65
+ "std": 0.4095070575542349
66
+ },
67
+ "avg_sample_grad_norm": {
68
+ "mean": 0.5011302649974823,
69
+ "std": 0.07381573007213547
70
+ },
71
+ "std_sample_grad_norm": {
72
+ "mean": 0.09378506834901998,
73
+ "std": 0.023664499864836274
74
+ },
75
+ "avg_cosine_similarity_to_mean": {
76
+ "mean": 0.5689752489328385,
77
+ "std": 0.030087469282562274
78
+ }
79
+ }
80
+ }
outputs/variance_comparison_all.pdf ADDED
Binary file (30.2 kB). View file
 
outputs/variance_comparison_all.png ADDED

Git LFS Details

  • SHA256: dcfe629ab5cc2dcf4604cfb0ec0b833a11b820043260cae1f9eafb1be71ba83d
  • Pointer size: 131 Bytes
  • Size of remote file: 517 kB
plot_all_variance.py ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Plot variance analysis results for SmolLM (Math), Qwen3 (Math), Maze."""
2
+
3
+ import matplotlib
4
+ matplotlib.use("Agg")
5
+ import matplotlib.pyplot as plt
6
+ import numpy as np
7
+ import os
8
+
9
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
10
+ OUTPUT_DIR = os.path.join(SCRIPT_DIR, "outputs")
11
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
12
+
13
+ # ── Data ──────────────────────────────────────────────────────────────────────
14
+ rollout_nums = [4, 8, 16, 32, 64, 128]
15
+
16
+ smollm_math = {
17
+ "blTrue": [3.461569e-01, 3.488942e-01, 2.763410e-01, 2.600237e-01, 2.039304e-01, 1.596079e-01],
18
+ "blFalse": [5.613685e-01, 4.680250e-01, 3.457040e-01, 2.965937e-01, 2.275286e-01, 1.723276e-01],
19
+ }
20
+
21
+ qwen3_math = {
22
+ "blTrue": [1.544762e-01, 1.679264e-01, 2.075920e-01, 1.788574e-01, 1.592376e-01, 1.314381e-01],
23
+ "blFalse": [2.140592e-01, 2.190761e-01, 2.448343e-01, 2.002312e-01, 1.702958e-01, 1.390878e-01],
24
+ }
25
+
26
+ maze = {
27
+ "blTrue": [9.343412e-02, 7.531368e-02, 5.179188e-02, 4.027390e-02, 2.930888e-02, 2.354821e-02],
28
+ "blFalse": [1.171640e-01, 8.123477e-02, 5.614680e-02, 4.236686e-02, 3.015591e-02, 2.521931e-02],
29
+ }
30
+
31
+ datasets = [
32
+ ("SmolLM-360M (GSM8k)", smollm_math),
33
+ ("Qwen3-1.7B (Polaris-53K)", qwen3_math),
34
+ ("Qwen2-3M (Maze)", maze),
35
+ ]
36
+
37
+ # ── Style ─────────────────────────────────────────────────────────────────────
38
+ plt.rcParams.update({
39
+ "font.size": 15,
40
+ "axes.titlesize": 18,
41
+ "axes.labelsize": 16,
42
+ "legend.fontsize": 14,
43
+ "xtick.labelsize": 14,
44
+ "ytick.labelsize": 14,
45
+ "figure.dpi": 150,
46
+ "savefig.dpi": 300,
47
+ "font.family": "sans-serif",
48
+ })
49
+
50
+ RED = "#D32F2F"
51
+ AMBER = "#F9A825"
52
+
53
+ # ── Plot ──────────────────────────────────────────────────────────────────────
54
+ fig, axes = plt.subplots(1, 3, figsize=(21, 6))
55
+
56
+ for ax, (title, data) in zip(axes, datasets):
57
+ xs = np.array(rollout_nums)
58
+ bl_true = np.array([v if v is not None else np.nan for v in data["blTrue"]])
59
+ bl_false = np.array([v if v is not None else np.nan for v in data["blFalse"]])
60
+
61
+ ax.plot(xs, bl_true, color=RED, marker="*", markersize=18, linewidth=4,
62
+ label="MaxRL", zorder=5)
63
+ ax.plot(xs, bl_false, color=AMBER, marker="o", markersize=11, linewidth=4,
64
+ label="MaxRL (w/o baseline)", zorder=4)
65
+
66
+ ax.set_xscale("log", base=2)
67
+ ax.set_xticks(rollout_nums)
68
+ ax.set_xticklabels([str(n) for n in rollout_nums])
69
+ ax.set_xlabel("Number of Rollouts", fontsize=16)
70
+ ax.set_ylabel("Gradient Variance", fontsize=16)
71
+ ax.set_title(title, fontsize=18, fontweight="bold", pad=12)
72
+ ax.legend(loc="upper right", framealpha=0.9, edgecolor="gray")
73
+ ax.grid(True, alpha=0.3, linestyle="--")
74
+ ax.spines["top"].set_visible(False)
75
+ ax.spines["right"].set_visible(False)
76
+
77
+ plt.tight_layout(w_pad=3)
78
+ out_path = os.path.join(OUTPUT_DIR, "variance_comparison_all.png")
79
+ plt.savefig(out_path, bbox_inches="tight")
80
+ plt.savefig(out_path.replace(".png", ".pdf"), bbox_inches="tight")
81
+ print(f"Saved to {out_path}")
plot_variance.py ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import matplotlib.pyplot as plt
3
+ import numpy as np
4
+
5
+ rollout_nums = [4, 8, 16, 32, 64, 128]
6
+
7
+ trace_var_bl_true = []
8
+ trace_var_bl_false = []
9
+
10
+ for nr in rollout_nums:
11
+ with open(f"results_bs16_nr{nr}_nb4_r5_blTrue.json") as f:
12
+ data = json.load(f)
13
+ trace_var_bl_true.append(data["averaged"]["trace_variance"]["mean"])
14
+
15
+ with open(f"results_bs16_nr{nr}_nb4_r5_blFalse.json") as f:
16
+ data = json.load(f)
17
+ trace_var_bl_false.append(data["averaged"]["trace_variance"]["mean"])
18
+
19
+ fig, ax = plt.subplots(figsize=(7, 5))
20
+
21
+ ax.plot(rollout_nums, trace_var_bl_true, marker='o', label='MaxRL')
22
+ ax.plot(rollout_nums, trace_var_bl_false, marker='s', label='MaxRL (w/o baseline)')
23
+
24
+ ax.set_xscale('log', base=2)
25
+ ax.set_xticks(rollout_nums)
26
+ ax.set_xticklabels(rollout_nums)
27
+ ax.set_xlabel('Rollout', fontsize=14)
28
+ ax.set_ylabel('Gradient Variance', fontsize=14)
29
+ ax.legend(fontsize=12)
30
+ ax.grid(True, alpha=0.3)
31
+
32
+ plt.tight_layout()
33
+ plt.savefig("variance_plot.pdf", dpi=300)
34
+ plt.savefig("variance_plot.png", dpi=300)
35
+ print("Saved variance_plot.pdf and variance_plot.png")
36
+
37
+ print("\nData:")
38
+ for nr, v_t, v_f in zip(rollout_nums, trace_var_bl_true, trace_var_bl_false):
39
+ print(f" nr={nr}: blTrue={v_t:.4f}, blFalse={v_f:.4f}")
qwen3_variance_analysis_auto.py ADDED
@@ -0,0 +1,400 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Automated Qwen3-1.7B Variance Analysis for MaxRL Policy Gradient.
3
+
4
+ Runs all 12 experiments (6 rollout_nums x 2 baseline settings) across multiple
5
+ GPUs with dynamic scheduling. Each GPU worker loads the model once and pulls
6
+ experiments from a shared task pool. Saves only trace_covariance mean/std to
7
+ outputs/ and plots the variance line chart.
8
+ """
9
+
10
+ import json
11
+ import os
12
+ import random
13
+ from functools import partial
14
+
15
+ import matplotlib
16
+ matplotlib.use("Agg")
17
+ import matplotlib.pyplot as plt
18
+ import numpy as np
19
+ import torch
20
+ import torch.multiprocessing as mp
21
+ from transformers import AutoModelForCausalLM, AutoTokenizer
22
+
23
+ # ============================================================================
24
+ # Configuration
25
+ # ============================================================================
26
+ BATCH_SIZE = 16
27
+ ROLLOUT_NUMS = [4, 8, 16, 32, 64, 128]
28
+ NUMBER_BATCHES_PER_ROUND = 4
29
+ TOTAL_ROUNDS = 32
30
+ MICRO_BATCH_SIZE = 2
31
+ MAX_SEQ_LEN = 4096
32
+ SEED = 42
33
+
34
+ MODEL_PATH = "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/qwen3_1.7B_Base_MaxRL_Polaris_1000_steps"
35
+ DATA_PATH = "/work/nvme/bgif/gzeng/MAXRL/variance_analysis/data/Qwen3/1.7b_512x512.jsonl"
36
+
37
+ GPU_IDS = [0, 1, 2, 3]
38
+ DTYPE = torch.bfloat16
39
+
40
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
41
+ OUTPUT_DIR = os.path.join(SCRIPT_DIR, "outputs", "Qwen3")
42
+
43
+
44
+ # ============================================================================
45
+ # Per-worker global state (initialized once per GPU worker)
46
+ # ============================================================================
47
+ _worker_model = None
48
+ _worker_tokenizer = None
49
+ _worker_prompt_data = None
50
+ _worker_all_prompt_ids = None
51
+ _worker_total_params = None
52
+ _worker_device = None
53
+
54
+
55
+ def worker_init(gpu_queue: mp.Queue):
56
+ """Called once per pool worker. Grabs a GPU and loads model + data."""
57
+ global _worker_model, _worker_tokenizer, _worker_prompt_data
58
+ global _worker_all_prompt_ids, _worker_total_params, _worker_device
59
+
60
+ gpu_id = gpu_queue.get()
61
+ _worker_device = f"cuda:{gpu_id}"
62
+ print(f"[Worker PID={os.getpid()}] Assigned to GPU {gpu_id}")
63
+
64
+ _worker_tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
65
+ if _worker_tokenizer.pad_token is None:
66
+ _worker_tokenizer.pad_token = _worker_tokenizer.eos_token
67
+
68
+ _worker_model = AutoModelForCausalLM.from_pretrained(
69
+ MODEL_PATH, torch_dtype=DTYPE,
70
+ ).to(_worker_device)
71
+ _worker_model.eval()
72
+ for p in _worker_model.parameters():
73
+ p.requires_grad_(True)
74
+
75
+ _worker_total_params = sum(p.numel() for p in _worker_model.parameters())
76
+ print(f"[GPU {gpu_id}] Model loaded: {_worker_total_params:,} parameters")
77
+
78
+ _worker_prompt_data = load_rollout_data(DATA_PATH)
79
+ _worker_all_prompt_ids = list(_worker_prompt_data.keys())
80
+
81
+
82
+ # ============================================================================
83
+ # Data Loading
84
+ # ============================================================================
85
+ def load_rollout_data(data_path: str) -> dict:
86
+ prompt_to_id = {}
87
+ prompt_data = {}
88
+
89
+ with open(data_path, "r") as f:
90
+ for line in f:
91
+ item = json.loads(line)
92
+ prompt_text = item["input"]
93
+ if prompt_text not in prompt_to_id:
94
+ pid = len(prompt_to_id)
95
+ prompt_to_id[prompt_text] = pid
96
+ prompt_data[pid] = {"input": prompt_text, "rollouts": []}
97
+ pid = prompt_to_id[prompt_text]
98
+ prompt_data[pid]["rollouts"].append({
99
+ "output": item["output"],
100
+ "score": item["score"],
101
+ })
102
+
103
+ print(f"Loaded {len(prompt_data)} prompts, "
104
+ f"each with {len(prompt_data[0]['rollouts'])} rollouts")
105
+ return prompt_data
106
+
107
+
108
+ # ============================================================================
109
+ # MaxRL Advantage Computation
110
+ # ============================================================================
111
+ def compute_maxrl_advantage(
112
+ scores: list[float], baseline: bool, epsilon: float = 1e-6,
113
+ ) -> list[float]:
114
+ mean = sum(scores) / len(scores)
115
+ if baseline:
116
+ return [(s - mean) / (mean + epsilon) for s in scores]
117
+ else:
118
+ return [s / (mean + epsilon) for s in scores]
119
+
120
+
121
+ # ============================================================================
122
+ # Tokenization & Batching
123
+ # ============================================================================
124
+ def tokenize_and_get_response_mask(
125
+ tokenizer, prompt: str, response: str, max_seq_len: int,
126
+ ) -> tuple[torch.Tensor, torch.Tensor]:
127
+ prompt_ids = tokenizer.encode(prompt, add_special_tokens=False)
128
+ response_ids = tokenizer.encode(response, add_special_tokens=False)
129
+
130
+ total_len = len(prompt_ids) + len(response_ids)
131
+ if total_len > max_seq_len:
132
+ max_resp = max_seq_len - len(prompt_ids)
133
+ if max_resp <= 0:
134
+ prompt_ids = prompt_ids[:max_seq_len // 2]
135
+ max_resp = max_seq_len - len(prompt_ids)
136
+ response_ids = response_ids[:max_resp]
137
+
138
+ input_ids = prompt_ids + response_ids
139
+ response_mask = [0] * len(prompt_ids) + [1] * len(response_ids)
140
+
141
+ return (
142
+ torch.tensor(input_ids, dtype=torch.long),
143
+ torch.tensor(response_mask, dtype=torch.float32),
144
+ )
145
+
146
+
147
+ def pad_batch(
148
+ batch_input_ids: list[torch.Tensor],
149
+ batch_response_masks: list[torch.Tensor],
150
+ pad_token_id: int,
151
+ ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
152
+ max_len = max(ids.shape[0] for ids in batch_input_ids)
153
+ B = len(batch_input_ids)
154
+
155
+ input_ids = torch.full((B, max_len), pad_token_id, dtype=torch.long)
156
+ response_mask = torch.zeros(B, max_len)
157
+ attention_mask = torch.zeros(B, max_len)
158
+
159
+ for i, (ids, rmask) in enumerate(zip(batch_input_ids, batch_response_masks)):
160
+ seq_len = ids.shape[0]
161
+ input_ids[i, max_len - seq_len:] = ids
162
+ response_mask[i, max_len - seq_len:] = rmask
163
+ attention_mask[i, max_len - seq_len:] = 1.0
164
+
165
+ return input_ids, response_mask, attention_mask
166
+
167
+
168
+ # ============================================================================
169
+ # Policy Gradient Loss
170
+ # ============================================================================
171
+ def compute_policy_gradient_loss(
172
+ model, input_ids, attention_mask, response_mask, advantages,
173
+ ) -> tuple[torch.Tensor, int]:
174
+ outputs = model(input_ids=input_ids, attention_mask=attention_mask)
175
+ logits = outputs.logits
176
+
177
+ shift_logits = logits[:, :-1, :]
178
+ shift_labels = input_ids[:, 1:]
179
+ shift_response_mask = response_mask[:, 1:]
180
+
181
+ log_probs = torch.log_softmax(shift_logits, dim=-1)
182
+ token_log_probs = torch.gather(
183
+ log_probs, dim=-1, index=shift_labels.unsqueeze(-1),
184
+ ).squeeze(-1)
185
+
186
+ token_losses = -advantages.unsqueeze(-1) * token_log_probs * shift_response_mask
187
+ valid_token_count = int(shift_response_mask.sum().item())
188
+ loss = token_losses.sum() / max(valid_token_count, 1)
189
+ return loss, valid_token_count
190
+
191
+
192
+ # ============================================================================
193
+ # Gradient Utilities
194
+ # ============================================================================
195
+ def collect_flat_gradient(model) -> torch.Tensor:
196
+ grads = []
197
+ for p in model.parameters():
198
+ if p.grad is not None:
199
+ grads.append(p.grad.detach().float().flatten())
200
+ else:
201
+ grads.append(torch.zeros(p.numel(), dtype=torch.float32, device=p.device))
202
+ return torch.cat(grads)
203
+
204
+
205
+ def compute_trace_variance(
206
+ grad_sum: torch.Tensor, grad_sq_sum: torch.Tensor, K: int,
207
+ ) -> float:
208
+ grad_mean = grad_sum / K
209
+ elementwise_var = (grad_sq_sum / K - grad_mean ** 2) * (K / (K - 1))
210
+ return elementwise_var.sum().item()
211
+
212
+
213
+ # ============================================================================
214
+ # Single Experiment (runs inside a worker process)
215
+ # ============================================================================
216
+ def run_single_experiment(task: tuple[int, bool]) -> tuple[str, dict]:
217
+ """Run one experiment using the worker's pre-loaded model and data.
218
+
219
+ Args:
220
+ task: (rollout_num, baseline)
221
+
222
+ Returns:
223
+ (key, {"mean": float, "std": float})
224
+ """
225
+ rollout_num, baseline = task
226
+ key = f"nr{rollout_num}_bl{baseline}"
227
+
228
+ model = _worker_model
229
+ tokenizer = _worker_tokenizer
230
+ prompt_data = _worker_prompt_data
231
+ all_prompt_ids = _worker_all_prompt_ids
232
+ total_params = _worker_total_params
233
+ device = _worker_device
234
+
235
+ print(f"[{device}] Starting {key}")
236
+
237
+ random.seed(SEED)
238
+ np.random.seed(SEED)
239
+ torch.manual_seed(SEED)
240
+
241
+ trace_variances = []
242
+
243
+ for round_idx in range(TOTAL_ROUNDS):
244
+ sampled_prompts = random.sample(all_prompt_ids, BATCH_SIZE)
245
+
246
+ rollouts_needed = NUMBER_BATCHES_PER_ROUND * rollout_num
247
+ round_rollout_subsets = {}
248
+ for pid in sampled_prompts:
249
+ rollouts = prompt_data[pid]["rollouts"]
250
+ if len(rollouts) < rollouts_needed:
251
+ raise ValueError(
252
+ f"Prompt {pid} has {len(rollouts)} rollouts, need {rollouts_needed}"
253
+ )
254
+ sampled = random.sample(rollouts, rollouts_needed)
255
+ round_rollout_subsets[pid] = [
256
+ sampled[s:s + rollout_num]
257
+ for s in range(0, rollouts_needed, rollout_num)
258
+ ]
259
+
260
+ grad_sum = torch.zeros(total_params, dtype=torch.float32)
261
+ grad_sq_sum = torch.zeros(total_params, dtype=torch.float32)
262
+
263
+ for subset_idx in range(NUMBER_BATCHES_PER_ROUND):
264
+ all_input_ids = []
265
+ all_response_masks = []
266
+ all_advantages = []
267
+
268
+ for pid in sampled_prompts:
269
+ prompt_text = prompt_data[pid]["input"]
270
+ sampled_rollouts = round_rollout_subsets[pid][subset_idx]
271
+ scores = [r["score"] for r in sampled_rollouts]
272
+ advantages = compute_maxrl_advantage(scores, baseline)
273
+
274
+ for rollout, adv in zip(sampled_rollouts, advantages):
275
+ ids, rmask = tokenize_and_get_response_mask(
276
+ tokenizer, prompt_text, rollout["output"], MAX_SEQ_LEN,
277
+ )
278
+ all_input_ids.append(ids)
279
+ all_response_masks.append(rmask)
280
+ all_advantages.append(adv)
281
+
282
+ model.zero_grad()
283
+ total_valid_tokens = int(
284
+ sum(rmask[1:].sum().item() for rmask in all_response_masks)
285
+ )
286
+ num_samples = len(all_input_ids)
287
+
288
+ for mb_start in range(0, num_samples, MICRO_BATCH_SIZE):
289
+ mb_end = min(mb_start + MICRO_BATCH_SIZE, num_samples)
290
+
291
+ mb_ids = all_input_ids[mb_start:mb_end]
292
+ mb_masks = all_response_masks[mb_start:mb_end]
293
+ mb_advs = all_advantages[mb_start:mb_end]
294
+
295
+ input_ids, response_mask, attention_mask = pad_batch(
296
+ mb_ids, mb_masks, tokenizer.pad_token_id,
297
+ )
298
+ input_ids = input_ids.to(device)
299
+ response_mask = response_mask.to(device)
300
+ attention_mask = attention_mask.to(device)
301
+ advantages_t = torch.tensor(mb_advs, dtype=DTYPE, device=device)
302
+
303
+ mb_loss, mb_valid_tokens = compute_policy_gradient_loss(
304
+ model, input_ids, attention_mask, response_mask, advantages_t,
305
+ )
306
+ scaled_loss = mb_loss * (mb_valid_tokens / max(total_valid_tokens, 1))
307
+ scaled_loss.backward()
308
+
309
+ flat_grad = collect_flat_gradient(model).cpu()
310
+ grad_sum += flat_grad
311
+ grad_sq_sum += flat_grad ** 2
312
+
313
+ trace_var = compute_trace_variance(
314
+ grad_sum, grad_sq_sum, NUMBER_BATCHES_PER_ROUND,
315
+ )
316
+ trace_variances.append(trace_var)
317
+ print(f" [{device}] {key} round {round_idx+1}/{TOTAL_ROUNDS}: "
318
+ f"trace_cov={trace_var:.6e}")
319
+
320
+ result = {
321
+ "mean": float(np.mean(trace_variances)),
322
+ "std": float(np.std(trace_variances)),
323
+ }
324
+ print(f"[{device}] Finished {key}: mean={result['mean']:.6e}, std={result['std']:.6e}")
325
+ return key, result
326
+
327
+
328
+ # ============================================================================
329
+ # Plotting
330
+ # ============================================================================
331
+ def plot_results(results: dict, output_dir: str):
332
+ rollout_nums = ROLLOUT_NUMS
333
+
334
+ means_bl_true = [results[f"nr{nr}_blTrue"]["mean"] for nr in rollout_nums]
335
+ means_bl_false = [results[f"nr{nr}_blFalse"]["mean"] for nr in rollout_nums]
336
+
337
+ fig, ax = plt.subplots(figsize=(7, 5))
338
+
339
+ ax.plot(rollout_nums, means_bl_true, marker='o', label='MaxRL')
340
+ ax.plot(rollout_nums, means_bl_false, marker='s', label='MaxRL (w/o baseline)')
341
+
342
+ ax.set_xscale('log', base=2)
343
+ ax.set_xticks(rollout_nums)
344
+ ax.set_xticklabels(rollout_nums)
345
+ ax.set_xlabel('Rollout', fontsize=14)
346
+ ax.set_ylabel('Gradient Variance', fontsize=14)
347
+ ax.legend(fontsize=12)
348
+ ax.grid(True, alpha=0.3)
349
+
350
+ plt.tight_layout()
351
+ plt.savefig(os.path.join(output_dir, "variance_plot.pdf"), dpi=300)
352
+ plt.savefig(os.path.join(output_dir, "variance_plot.png"), dpi=300)
353
+ print(f"Plots saved to {output_dir}/variance_plot.{{pdf,png}}")
354
+
355
+
356
+ # ============================================================================
357
+ # Main
358
+ # ============================================================================
359
+ def main():
360
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
361
+
362
+ # Build task list: 12 experiments
363
+ tasks = []
364
+ for rollout_num in ROLLOUT_NUMS:
365
+ for baseline in [True, False]:
366
+ tasks.append((rollout_num, baseline))
367
+
368
+ print(f"Scheduling {len(tasks)} experiments across {len(GPU_IDS)} GPUs")
369
+
370
+ # GPU queue: each worker grabs one GPU ID on init
371
+ gpu_queue = mp.Queue()
372
+ for gid in GPU_IDS:
373
+ gpu_queue.put(gid)
374
+
375
+ # Pool of workers = number of GPUs. Each worker inits once (loads model),
376
+ # then processes tasks dynamically from the pool.
377
+ with mp.Pool(
378
+ processes=len(GPU_IDS),
379
+ initializer=worker_init,
380
+ initargs=(gpu_queue,),
381
+ ) as pool:
382
+ results_list = pool.map(run_single_experiment, tasks)
383
+
384
+ # Collect results
385
+ results = dict(results_list)
386
+
387
+ # Save
388
+ results_path = os.path.join(OUTPUT_DIR, "results.json")
389
+ with open(results_path, "w") as f:
390
+ json.dump(results, f, indent=2)
391
+ print(f"Results saved to {results_path}")
392
+
393
+ # Plot
394
+ plot_results(results, OUTPUT_DIR)
395
+ print("All done!")
396
+
397
+
398
+ if __name__ == "__main__":
399
+ mp.set_start_method("spawn")
400
+ main()
qwen3_variance_analysis_resume.py ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Resume script for Qwen3 variance analysis.
3
+ Only runs the 2 experiments interrupted by job 2033368:
4
+ nr128_blTrue, nr128_blFalse
5
+ Then merges with the 10 completed results from the killed run and saves/plots.
6
+ """
7
+
8
+ import json
9
+ import os
10
+ import sys
11
+
12
+ # Reuse everything from the original script
13
+ sys.path.insert(0, os.path.dirname(os.path.abspath(__file__)))
14
+ from qwen3_variance_analysis_auto import (
15
+ ROLLOUT_NUMS, OUTPUT_DIR,
16
+ worker_init, run_single_experiment, plot_results,
17
+ )
18
+ import torch.multiprocessing as mp
19
+
20
+
21
+ # Results already completed in job 2033368 (extracted from log)
22
+ COMPLETED_RESULTS = {
23
+ "nr4_blFalse": {"mean": 2.453292e-01, "std": 8.433693e-02},
24
+ "nr4_blTrue": {"mean": 1.544762e-01, "std": 5.542017e-02},
25
+ "nr8_blTrue": {"mean": 1.679264e-01, "std": 5.820824e-02},
26
+ "nr8_blFalse": {"mean": 2.190761e-01, "std": 7.012213e-02},
27
+ "nr16_blFalse": {"mean": 2.448343e-01, "std": 7.550860e-02},
28
+ "nr16_blTrue": {"mean": 2.075920e-01, "std": 6.842490e-02},
29
+ "nr32_blTrue": {"mean": 1.788574e-01, "std": 5.623576e-02},
30
+ "nr32_blFalse": {"mean": 2.002312e-01, "std": 5.805813e-02},
31
+ "nr64_blFalse": {"mean": 1.702958e-01, "std": 5.899355e-02},
32
+ "nr64_blTrue": {"mean": 1.592376e-01, "std": 5.725100e-02},
33
+ }
34
+
35
+ # Only need to run these 2
36
+ REMAINING_TASKS = [
37
+ (128, True), # nr128_blTrue
38
+ (128, False), # nr128_blFalse
39
+ ]
40
+
41
+
42
+ def main():
43
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
44
+
45
+ print(f"Resuming: {len(REMAINING_TASKS)} experiments on 2 GPUs")
46
+
47
+ gpu_queue = mp.Queue()
48
+ for gid in [0, 1]:
49
+ gpu_queue.put(gid)
50
+
51
+ with mp.Pool(
52
+ processes=2,
53
+ initializer=worker_init,
54
+ initargs=(gpu_queue,),
55
+ ) as pool:
56
+ new_results_list = pool.map(run_single_experiment, REMAINING_TASKS)
57
+
58
+ # Merge
59
+ results = dict(COMPLETED_RESULTS)
60
+ results.update(dict(new_results_list))
61
+
62
+ # Save
63
+ results_path = os.path.join(OUTPUT_DIR, "results.json")
64
+ with open(results_path, "w") as f:
65
+ json.dump(results, f, indent=2)
66
+ print(f"Results saved to {results_path}")
67
+
68
+ # Plot
69
+ plot_results(results, OUTPUT_DIR)
70
+ print("All done!")
71
+
72
+
73
+ if __name__ == "__main__":
74
+ mp.set_start_method("spawn")
75
+ main()
run_maze_auto.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=var_maze
3
+ #SBATCH --nodes=1
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --cpus-per-task=64
6
+ #SBATCH --gres=gpu:4
7
+ #SBATCH --time=24:00:00
8
+ #SBATCH --mem=256G
9
+ #SBATCH --output=logs/%x_%j.log
10
+ #SBATCH --partition=ghx4
11
+ #SBATCH --account=bgif-dtai-gh
12
+ #SBATCH --reservation=sup-24244
13
+
14
+ source /u/gzeng/.bashrc >/dev/null 2>&1 || true
15
+ source /u/gzeng/miniconda3/etc/profile.d/conda.sh
16
+ conda activate exploration
17
+
18
+ set -Eeuo pipefail
19
+
20
+ export CACHE="/work/nvme/bgif/gzeng/MAXRL/cache"
21
+ export HF_HOME="${CACHE}/huggingface"
22
+ export HUGGINGFACE_HUB_CACHE="${HF_HOME}/hub"
23
+ unset TRANSFORMERS_CACHE
24
+
25
+ export PYTHONUNBUFFERED=1
26
+ export NCCL_DEBUG=WARN
27
+ export TOKENIZERS_PARALLELISM=false
28
+
29
+ mkdir -p logs
30
+
31
+ echo "Start: $(date -Ins)"
32
+ echo "Node: $(hostname)"
33
+ echo "GPUs: $(nvidia-smi -L 2>/dev/null | wc -l)"
34
+
35
+ cd /work/nvme/bgif/gzeng/MAXRL/variance_analysis
36
+ python maze_variance_analysis_auto.py
37
+
38
+ echo "Done: $(date -Ins)"
run_qwen3_auto.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=var_qwen3
3
+ #SBATCH --nodes=1
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --cpus-per-task=64
6
+ #SBATCH --gres=gpu:4
7
+ #SBATCH --time=48:00:00
8
+ #SBATCH --mem=256G
9
+ #SBATCH --output=logs/%x_%j.log
10
+ #SBATCH --partition=ghx4
11
+ #SBATCH --account=bgif-dtai-gh
12
+ #SBATCH --reservation=sup-24244
13
+
14
+ source /u/gzeng/.bashrc >/dev/null 2>&1 || true
15
+ source /u/gzeng/miniconda3/etc/profile.d/conda.sh
16
+ conda activate exploration
17
+
18
+ set -Eeuo pipefail
19
+
20
+ export CACHE="/work/nvme/bgif/gzeng/MAXRL/cache"
21
+ export HF_HOME="${CACHE}/huggingface"
22
+ export HUGGINGFACE_HUB_CACHE="${HF_HOME}/hub"
23
+ unset TRANSFORMERS_CACHE
24
+
25
+ export PYTHONUNBUFFERED=1
26
+ export NCCL_DEBUG=WARN
27
+ export TOKENIZERS_PARALLELISM=false
28
+
29
+ mkdir -p logs
30
+
31
+ echo "Start: $(date -Ins)"
32
+ echo "Node: $(hostname)"
33
+ echo "GPUs: $(nvidia-smi -L 2>/dev/null | wc -l)"
34
+
35
+ cd /work/nvme/bgif/gzeng/MAXRL/variance_analysis
36
+ python qwen3_variance_analysis_auto.py
37
+
38
+ echo "Done: $(date -Ins)"
run_qwen3_resume.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=var_qwen3_resume
3
+ #SBATCH --nodes=1
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --cpus-per-task=128
6
+ #SBATCH --gres=gpu:4
7
+ #SBATCH --time=12:00:00
8
+ #SBATCH --mem=512G
9
+ #SBATCH --output=logs/%x_%j.log
10
+ #SBATCH --partition=ghx4
11
+ #SBATCH --account=bgif-dtai-gh
12
+ #SBATCH --reservation=sup-24244
13
+
14
+ source /u/gzeng/.bashrc >/dev/null 2>&1 || true
15
+ source /u/gzeng/miniconda3/etc/profile.d/conda.sh
16
+ conda activate exploration
17
+
18
+ set -Eeuo pipefail
19
+
20
+ export CACHE="/work/nvme/bgif/gzeng/MAXRL/cache"
21
+ export HF_HOME="${CACHE}/huggingface"
22
+ export HUGGINGFACE_HUB_CACHE="${HF_HOME}/hub"
23
+ unset TRANSFORMERS_CACHE
24
+
25
+ export PYTHONUNBUFFERED=1
26
+ export NCCL_DEBUG=WARN
27
+ export TOKENIZERS_PARALLELISM=false
28
+
29
+ mkdir -p logs
30
+
31
+ echo "Start: $(date -Ins)"
32
+ echo "Node: $(hostname)"
33
+ echo "GPUs: $(nvidia-smi -L 2>/dev/null | wc -l)"
34
+
35
+ cd /work/nvme/bgif/gzeng/MAXRL/variance_analysis
36
+ python qwen3_variance_analysis_resume.py
37
+
38
+ echo "Done: $(date -Ins)"
run_smollm_auto.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=var_smollm
3
+ #SBATCH --nodes=1
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --cpus-per-task=64
6
+ #SBATCH --gres=gpu:4
7
+ #SBATCH --time=24:00:00
8
+ #SBATCH --mem=256G
9
+ #SBATCH --output=logs/%x_%j.log
10
+ #SBATCH --partition=ghx4
11
+ #SBATCH --account=bgif-dtai-gh
12
+ #SBATCH --reservation=sup-24244
13
+
14
+ source /u/gzeng/.bashrc >/dev/null 2>&1 || true
15
+ source /u/gzeng/miniconda3/etc/profile.d/conda.sh
16
+ conda activate exploration
17
+
18
+ set -Eeuo pipefail
19
+
20
+ export CACHE="/work/nvme/bgif/gzeng/MAXRL/cache"
21
+ export HF_HOME="${CACHE}/huggingface"
22
+ export HUGGINGFACE_HUB_CACHE="${HF_HOME}/hub"
23
+ unset TRANSFORMERS_CACHE
24
+
25
+ export PYTHONUNBUFFERED=1
26
+ export NCCL_DEBUG=WARN
27
+ export TOKENIZERS_PARALLELISM=false
28
+
29
+ mkdir -p logs
30
+
31
+ echo "Start: $(date -Ins)"
32
+ echo "Node: $(hostname)"
33
+ echo "GPUs: $(nvidia-smi -L 2>/dev/null | wc -l)"
34
+
35
+ cd /work/nvme/bgif/gzeng/MAXRL/variance_analysis
36
+ python smollm_variance_analysis_auto.py
37
+
38
+ echo "Done: $(date -Ins)"
run_smollm_snr.sh ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+ #SBATCH --job-name=snr_smollm
3
+ #SBATCH --nodes=1
4
+ #SBATCH --ntasks-per-node=1
5
+ #SBATCH --cpus-per-task=64
6
+ #SBATCH --gres=gpu:4
7
+ #SBATCH --time=24:00:00
8
+ #SBATCH --mem=256G
9
+ #SBATCH --output=logs/%x_%j.log
10
+ #SBATCH --partition=ghx4
11
+ #SBATCH --account=bgif-dtai-gh
12
+ #SBATCH --reservation=sup-24244
13
+
14
+ source /u/gzeng/.bashrc >/dev/null 2>&1 || true
15
+ source /u/gzeng/miniconda3/etc/profile.d/conda.sh
16
+ conda activate exploration
17
+
18
+ set -Eeuo pipefail
19
+
20
+ export CACHE="/work/nvme/bgif/gzeng/MAXRL/cache"
21
+ export HF_HOME="${CACHE}/huggingface"
22
+ export HUGGINGFACE_HUB_CACHE="${HF_HOME}/hub"
23
+ unset TRANSFORMERS_CACHE
24
+
25
+ export PYTHONUNBUFFERED=1
26
+ export NCCL_DEBUG=WARN
27
+ export TOKENIZERS_PARALLELISM=false
28
+
29
+ mkdir -p logs
30
+
31
+ echo "Start: $(date -Ins)"
32
+ echo "Node: $(hostname)"
33
+ echo "GPUs: $(nvidia-smi -L 2>/dev/null | wc -l)"
34
+
35
+ cd /work/nvme/bgif/gzeng/MAXRL/variance_analysis
36
+ python smollm_snr_analysis_auto.py
37
+
38
+ echo "Done: $(date -Ins)"
smollm_snr_analysis_auto.py ADDED
@@ -0,0 +1,442 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ SmolLM Gradient SNR Analysis (Daman-style).
3
+
4
+ For a FIXED batch of 64 prompts, compute S=64 gradient samples per
5
+ (advantage_type, rollout_num) pair. Each gradient sample re-samples rollouts
6
+ from the pre-computed pool, so the only source of variance is rollout sampling.
7
+
8
+ Reports SNR = ||mean(grad)||^2 / sum(var(grad)) for MaxRL, GRPO, and RLOO.
9
+ Distributes experiments across GPUs with dynamic scheduling.
10
+ """
11
+
12
+ import json
13
+ import os
14
+ import random
15
+ from functools import partial
16
+
17
+ import matplotlib
18
+ matplotlib.use("Agg")
19
+ import matplotlib.pyplot as plt
20
+ import numpy as np
21
+ import torch
22
+ import torch.multiprocessing as mp
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+
25
+ # ============================================================================
26
+ # Configuration
27
+ # ============================================================================
28
+ BATCH_SIZE = 64 # fixed first 64 prompts
29
+ ROLLOUT_NUMS = [4, 8]
30
+ S = 64 # number of gradient samples for SNR estimation
31
+ MICRO_BATCH_SIZE = 8
32
+ MAX_SEQ_LEN = 2048
33
+ SEED = 42
34
+
35
+ ADVANTAGE_TYPES = ["maxrl", "grpo", "rloo"]
36
+
37
+ MODEL_PATH = "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps"
38
+ DATA_PATH = "/work/nvme/bgif/gzeng/MAXRL/variance_analysis/data/SmolLM/512x512.jsonl"
39
+
40
+ GPU_IDS = [0, 1, 2, 3]
41
+ DTYPE = torch.bfloat16
42
+
43
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
44
+ OUTPUT_DIR = os.path.join(SCRIPT_DIR, "outputs", "SmolLM_SNR")
45
+
46
+ # ============================================================================
47
+ # Per-worker global state
48
+ # ============================================================================
49
+ _worker_model = None
50
+ _worker_tokenizer = None
51
+ _worker_prompt_data = None
52
+ _worker_fixed_prompt_ids = None
53
+ _worker_total_params = None
54
+ _worker_device = None
55
+
56
+
57
+ def worker_init(gpu_queue: mp.Queue):
58
+ global _worker_model, _worker_tokenizer, _worker_prompt_data
59
+ global _worker_fixed_prompt_ids, _worker_total_params, _worker_device
60
+
61
+ gpu_id = gpu_queue.get()
62
+ _worker_device = f"cuda:{gpu_id}"
63
+ print(f"[Worker PID={os.getpid()}] Assigned to GPU {gpu_id}")
64
+
65
+ _worker_tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
66
+ if _worker_tokenizer.pad_token is None:
67
+ _worker_tokenizer.pad_token = _worker_tokenizer.eos_token
68
+
69
+ _worker_model = AutoModelForCausalLM.from_pretrained(
70
+ MODEL_PATH, torch_dtype=DTYPE,
71
+ ).to(_worker_device)
72
+ _worker_model.eval()
73
+ for p in _worker_model.parameters():
74
+ p.requires_grad_(True)
75
+
76
+ _worker_total_params = sum(p.numel() for p in _worker_model.parameters())
77
+ print(f"[GPU {gpu_id}] Model loaded: {_worker_total_params:,} parameters")
78
+
79
+ _worker_prompt_data = load_rollout_data(DATA_PATH)
80
+ # Fix the first BATCH_SIZE prompts
81
+ _worker_fixed_prompt_ids = list(range(BATCH_SIZE))
82
+
83
+
84
+ # ============================================================================
85
+ # Data Loading
86
+ # ============================================================================
87
+ def load_rollout_data(data_path: str) -> dict:
88
+ prompt_to_id = {}
89
+ prompt_data = {}
90
+
91
+ with open(data_path, "r") as f:
92
+ for line in f:
93
+ item = json.loads(line)
94
+ prompt_text = item["input"]
95
+ if prompt_text not in prompt_to_id:
96
+ pid = len(prompt_to_id)
97
+ prompt_to_id[prompt_text] = pid
98
+ prompt_data[pid] = {"input": prompt_text, "rollouts": []}
99
+ pid = prompt_to_id[prompt_text]
100
+ prompt_data[pid]["rollouts"].append({
101
+ "output": item["output"],
102
+ "score": item["score"],
103
+ })
104
+
105
+ print(f"Loaded {len(prompt_data)} prompts, "
106
+ f"each with {len(prompt_data[0]['rollouts'])} rollouts")
107
+ return prompt_data
108
+
109
+
110
+ # ============================================================================
111
+ # Advantage Computation
112
+ # ============================================================================
113
+ def compute_advantage(scores: list[float], advantage_type: str, epsilon: float = 1e-6) -> list[float]:
114
+ n = len(scores)
115
+ mean = sum(scores) / n
116
+
117
+ if advantage_type == "maxrl":
118
+ # (score - mean) / (mean + eps)
119
+ return [(s - mean) / (mean + epsilon) for s in scores]
120
+
121
+ elif advantage_type == "grpo":
122
+ # (score - mean) / (std + eps)
123
+ var = sum((s - mean) ** 2 for s in scores) / n
124
+ std = var ** 0.5
125
+ return [(s - mean) / (std + epsilon) for s in scores]
126
+
127
+ elif advantage_type == "rloo":
128
+ # REINFORCE Leave-One-Out: advantage_i = score_i - mean_{j != i}
129
+ total = sum(scores)
130
+ return [s - (total - s) / (n - 1) for s in scores]
131
+
132
+ else:
133
+ raise ValueError(f"Unknown advantage type: {advantage_type}")
134
+
135
+
136
+ # ============================================================================
137
+ # Tokenization & Batching
138
+ # ============================================================================
139
+ def tokenize_and_get_response_mask(
140
+ tokenizer, prompt: str, response: str, max_seq_len: int,
141
+ ) -> tuple[torch.Tensor, torch.Tensor]:
142
+ prompt_ids = tokenizer.encode(prompt, add_special_tokens=False)
143
+ response_ids = tokenizer.encode(response, add_special_tokens=False)
144
+
145
+ total_len = len(prompt_ids) + len(response_ids)
146
+ if total_len > max_seq_len:
147
+ max_resp = max_seq_len - len(prompt_ids)
148
+ if max_resp <= 0:
149
+ prompt_ids = prompt_ids[:max_seq_len // 2]
150
+ max_resp = max_seq_len - len(prompt_ids)
151
+ response_ids = response_ids[:max_resp]
152
+
153
+ input_ids = prompt_ids + response_ids
154
+ response_mask = [0] * len(prompt_ids) + [1] * len(response_ids)
155
+
156
+ return (
157
+ torch.tensor(input_ids, dtype=torch.long),
158
+ torch.tensor(response_mask, dtype=torch.float32),
159
+ )
160
+
161
+
162
+ def pad_batch(
163
+ batch_input_ids: list[torch.Tensor],
164
+ batch_response_masks: list[torch.Tensor],
165
+ pad_token_id: int,
166
+ ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
167
+ max_len = max(ids.shape[0] for ids in batch_input_ids)
168
+ B = len(batch_input_ids)
169
+
170
+ input_ids = torch.full((B, max_len), pad_token_id, dtype=torch.long)
171
+ response_mask = torch.zeros(B, max_len)
172
+ attention_mask = torch.zeros(B, max_len)
173
+
174
+ for i, (ids, rmask) in enumerate(zip(batch_input_ids, batch_response_masks)):
175
+ seq_len = ids.shape[0]
176
+ input_ids[i, max_len - seq_len:] = ids
177
+ response_mask[i, max_len - seq_len:] = rmask
178
+ attention_mask[i, max_len - seq_len:] = 1.0
179
+
180
+ return input_ids, response_mask, attention_mask
181
+
182
+
183
+ # ============================================================================
184
+ # Policy Gradient Loss
185
+ # ============================================================================
186
+ def compute_policy_gradient_loss(
187
+ model, input_ids, attention_mask, response_mask, advantages,
188
+ ) -> tuple[torch.Tensor, int]:
189
+ outputs = model(input_ids=input_ids, attention_mask=attention_mask)
190
+ logits = outputs.logits
191
+
192
+ shift_logits = logits[:, :-1, :]
193
+ shift_labels = input_ids[:, 1:]
194
+ shift_response_mask = response_mask[:, 1:]
195
+
196
+ log_probs = torch.log_softmax(shift_logits, dim=-1)
197
+ token_log_probs = torch.gather(
198
+ log_probs, dim=-1, index=shift_labels.unsqueeze(-1),
199
+ ).squeeze(-1)
200
+
201
+ token_losses = -advantages.unsqueeze(-1) * token_log_probs * shift_response_mask
202
+ valid_token_count = int(shift_response_mask.sum().item())
203
+ loss = token_losses.sum() / max(valid_token_count, 1)
204
+ return loss, valid_token_count
205
+
206
+
207
+ # ============================================================================
208
+ # Gradient Utilities
209
+ # ============================================================================
210
+ def collect_flat_gradient(model) -> torch.Tensor:
211
+ grads = []
212
+ for p in model.parameters():
213
+ if p.grad is not None:
214
+ grads.append(p.grad.detach().float().flatten())
215
+ else:
216
+ grads.append(torch.zeros(p.numel(), dtype=torch.float32, device=p.device))
217
+ return torch.cat(grads)
218
+
219
+
220
+ def gradient_snr(gradients: torch.Tensor, eps: float = 1e-8):
221
+ """
222
+ Compute gradient SNR from (S, D) tensor of gradient vectors.
223
+ Returns: (snr, mean_sq_norm, var_sum)
224
+ """
225
+ mu = gradients.mean(dim=0)
226
+ var = gradients.var(dim=0, unbiased=False)
227
+ mean_sq_norm = mu.pow(2).sum()
228
+ var_sum = var.sum()
229
+ snr = mean_sq_norm / (var_sum + eps)
230
+ return snr.item(), mean_sq_norm.item(), var_sum.item()
231
+
232
+
233
+ # ============================================================================
234
+ # Single Experiment
235
+ # ============================================================================
236
+ def run_single_experiment(task: tuple[str, int]) -> tuple[str, dict]:
237
+ advantage_type, rollout_num = task
238
+ key = f"{advantage_type}_nr{rollout_num}"
239
+
240
+ model = _worker_model
241
+ tokenizer = _worker_tokenizer
242
+ prompt_data = _worker_prompt_data
243
+ fixed_prompt_ids = _worker_fixed_prompt_ids
244
+ device = _worker_device
245
+
246
+ print(f"[{device}] Starting {key}")
247
+
248
+ all_grads = []
249
+
250
+ for s in range(S):
251
+ random.seed(SEED + s)
252
+
253
+ # For each prompt, sample rollout_num rollouts from the pool
254
+ all_input_ids = []
255
+ all_response_masks = []
256
+ all_advantages = []
257
+
258
+ for pid in fixed_prompt_ids:
259
+ rollouts = prompt_data[pid]["rollouts"]
260
+ sampled = random.sample(rollouts, rollout_num)
261
+ scores = [r["score"] for r in sampled]
262
+ advantages = compute_advantage(scores, advantage_type)
263
+
264
+ for rollout, adv in zip(sampled, advantages):
265
+ ids, rmask = tokenize_and_get_response_mask(
266
+ tokenizer, prompt_data[pid]["input"], rollout["output"], MAX_SEQ_LEN,
267
+ )
268
+ all_input_ids.append(ids)
269
+ all_response_masks.append(rmask)
270
+ all_advantages.append(adv)
271
+
272
+ # Forward + backward with micro-batching
273
+ model.zero_grad()
274
+ total_valid_tokens = int(
275
+ sum(rmask[1:].sum().item() for rmask in all_response_masks)
276
+ )
277
+ num_samples = len(all_input_ids)
278
+
279
+ for mb_start in range(0, num_samples, MICRO_BATCH_SIZE):
280
+ mb_end = min(mb_start + MICRO_BATCH_SIZE, num_samples)
281
+
282
+ mb_ids = all_input_ids[mb_start:mb_end]
283
+ mb_masks = all_response_masks[mb_start:mb_end]
284
+ mb_advs = all_advantages[mb_start:mb_end]
285
+
286
+ input_ids, response_mask, attention_mask = pad_batch(
287
+ mb_ids, mb_masks, tokenizer.pad_token_id,
288
+ )
289
+ input_ids = input_ids.to(device)
290
+ response_mask = response_mask.to(device)
291
+ attention_mask = attention_mask.to(device)
292
+ advantages_t = torch.tensor(mb_advs, dtype=DTYPE, device=device)
293
+
294
+ mb_loss, mb_valid_tokens = compute_policy_gradient_loss(
295
+ model, input_ids, attention_mask, response_mask, advantages_t,
296
+ )
297
+ scaled_loss = mb_loss * (mb_valid_tokens / max(total_valid_tokens, 1))
298
+ scaled_loss.backward()
299
+
300
+ flat_grad = collect_flat_gradient(model).cpu()
301
+ all_grads.append(flat_grad)
302
+
303
+ if (s + 1) % 16 == 0:
304
+ print(f" [{device}] {key}: {s+1}/{S} gradient samples collected")
305
+
306
+ gradients = torch.stack(all_grads) # (S, D)
307
+ snr, mean_sq_norm, var_sum = gradient_snr(gradients)
308
+ print(f"[{device}] {key}: SNR={snr:.6f}, mean={mean_sq_norm:.6e}, var={var_sum:.6e}")
309
+
310
+ result = {
311
+ "snr": snr,
312
+ "mean": mean_sq_norm,
313
+ "var": var_sum,
314
+ }
315
+ return key, result
316
+
317
+
318
+ # ============================================================================
319
+ # Plotting
320
+ # ============================================================================
321
+ LABEL_MAP = {
322
+ "maxrl": "MaxRL",
323
+ "grpo": "GRPO",
324
+ "rloo": "RLOO",
325
+ }
326
+
327
+ COLOR_MAP = {
328
+ "maxrl": "#e74c3c",
329
+ "grpo": "#3498db",
330
+ "rloo": "#2ecc71",
331
+ }
332
+
333
+
334
+ def plot_results(results: dict, output_dir: str):
335
+ fig, axes = plt.subplots(1, 3, figsize=(18, 5))
336
+
337
+ # --- SNR ---
338
+ ax = axes[0]
339
+ for adv_type in ADVANTAGE_TYPES:
340
+ xs, ys = [], []
341
+ for nr in ROLLOUT_NUMS:
342
+ key = f"{adv_type}_nr{nr}"
343
+ if key in results and results[key] is not None:
344
+ xs.append(nr)
345
+ ys.append(results[key]["snr"])
346
+ label = LABEL_MAP.get(adv_type, adv_type)
347
+ ax.plot(xs, ys, marker="o", label=label, color=COLOR_MAP.get(adv_type))
348
+ ax.set_xscale("log", base=2)
349
+ ax.set_xticks(ROLLOUT_NUMS)
350
+ ax.set_xticklabels(ROLLOUT_NUMS)
351
+ ax.set_xlabel("Rollouts (N)")
352
+ ax.set_ylabel("Gradient SNR")
353
+ ax.set_title(f"Gradient SNR (S={S}, batch={BATCH_SIZE})")
354
+ ax.legend()
355
+ ax.grid(True, which="both", linestyle="--", alpha=0.5)
356
+
357
+ # --- Mean (signal) ---
358
+ ax = axes[1]
359
+ for adv_type in ADVANTAGE_TYPES:
360
+ xs, ys = [], []
361
+ for nr in ROLLOUT_NUMS:
362
+ key = f"{adv_type}_nr{nr}"
363
+ if key in results and results[key] is not None:
364
+ xs.append(nr)
365
+ ys.append(results[key]["mean"])
366
+ label = LABEL_MAP.get(adv_type, adv_type)
367
+ ax.plot(xs, ys, marker="o", label=label, color=COLOR_MAP.get(adv_type))
368
+ ax.set_xscale("log", base=2)
369
+ ax.set_xticks(ROLLOUT_NUMS)
370
+ ax.set_xticklabels(ROLLOUT_NUMS)
371
+ ax.set_xlabel("Rollouts (N)")
372
+ ax.set_ylabel("||mean(grad)||²")
373
+ ax.set_title("Signal (mean gradient norm²)")
374
+ ax.legend()
375
+ ax.grid(True, which="both", linestyle="--", alpha=0.5)
376
+
377
+ # --- Var (noise) ---
378
+ ax = axes[2]
379
+ for adv_type in ADVANTAGE_TYPES:
380
+ xs, ys = [], []
381
+ for nr in ROLLOUT_NUMS:
382
+ key = f"{adv_type}_nr{nr}"
383
+ if key in results and results[key] is not None:
384
+ xs.append(nr)
385
+ ys.append(results[key]["var"])
386
+ label = LABEL_MAP.get(adv_type, adv_type)
387
+ ax.plot(xs, ys, marker="o", label=label, color=COLOR_MAP.get(adv_type))
388
+ ax.set_xscale("log", base=2)
389
+ ax.set_xticks(ROLLOUT_NUMS)
390
+ ax.set_xticklabels(ROLLOUT_NUMS)
391
+ ax.set_xlabel("Rollouts (N)")
392
+ ax.set_ylabel("sum(var(grad))")
393
+ ax.set_title("Noise (gradient variance)")
394
+ ax.legend()
395
+ ax.grid(True, which="both", linestyle="--", alpha=0.5)
396
+
397
+ plt.tight_layout()
398
+ plt.savefig(os.path.join(output_dir, "snr_plot.pdf"), dpi=300)
399
+ plt.savefig(os.path.join(output_dir, "snr_plot.png"), dpi=300)
400
+ print(f"Plots saved to {output_dir}/snr_plot.{{pdf,png}}")
401
+
402
+
403
+ # ============================================================================
404
+ # Main
405
+ # ============================================================================
406
+ def main():
407
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
408
+
409
+ # Build task list: 3 advantage types x 6 rollout nums = 18 experiments
410
+ tasks = []
411
+ for adv_type in ADVANTAGE_TYPES:
412
+ for rollout_num in ROLLOUT_NUMS:
413
+ tasks.append((adv_type, rollout_num))
414
+
415
+ print(f"Scheduling {len(tasks)} experiments across {len(GPU_IDS)} GPUs")
416
+ print(f"Fixed batch: first {BATCH_SIZE} prompts, S={S} gradient samples each")
417
+
418
+ gpu_queue = mp.Queue()
419
+ for gid in GPU_IDS:
420
+ gpu_queue.put(gid)
421
+
422
+ with mp.Pool(
423
+ processes=len(GPU_IDS),
424
+ initializer=worker_init,
425
+ initargs=(gpu_queue,),
426
+ ) as pool:
427
+ results_list = pool.map(run_single_experiment, tasks)
428
+
429
+ results = dict(results_list)
430
+
431
+ results_path = os.path.join(OUTPUT_DIR, "snr_results.json")
432
+ with open(results_path, "w") as f:
433
+ json.dump(results, f, indent=2)
434
+ print(f"Results saved to {results_path}")
435
+
436
+ plot_results(results, OUTPUT_DIR)
437
+ print("All done!")
438
+
439
+
440
+ if __name__ == "__main__":
441
+ mp.set_start_method("spawn")
442
+ main()
smollm_variance_analysis.py ADDED
@@ -0,0 +1,478 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ SmolLM Variance Analysis for MaxRL Policy Gradient.
3
+
4
+ Measures the gradient variance of MaxRL's policy gradient estimator by sampling
5
+ different rollout subsets from pre-computed data and computing how much the
6
+ resulting policy gradients vary.
7
+
8
+ This script also supports an ablation on the MaxRL baseline term. Here,
9
+ `BASELINE=True` means we use the standard MaxRL-style mean-centering in the
10
+ numerator:
11
+
12
+ (score - mean_score) / (mean_score + epsilon)
13
+
14
+ and `BASELINE=False` removes only that centering term while keeping the MaxRL
15
+ normalization in the denominator:
16
+
17
+ score / (mean_score + epsilon)
18
+
19
+ So the ablation isolates the effect of the baseline term inside MaxRL rather
20
+ than switching to vanilla REINFORCE.
21
+
22
+ Within each round, rollout subsets for the same prompt are sampled without
23
+ replacement across subsets, so different subsets do not reuse the same rollout.
24
+ """
25
+
26
+ import json
27
+ import os
28
+ import random
29
+
30
+ import numpy as np
31
+ import torch
32
+ from transformers import AutoModelForCausalLM, AutoTokenizer
33
+
34
+ # ============================================================================
35
+ # Global Configuration
36
+ # ============================================================================
37
+ BATCH_SIZE = 16 # number of prompts per round
38
+ ROLLOUT_NUM = 4 # rollouts sampled per prompt per subset
39
+ NUMBER_BATCHES_PER_ROUND = 4 # number of different rollout subsets per round
40
+ TOTAL_ROUNDS = 5 # rounds to average over
41
+ BASELINE = False # if True: (score - mean)/(mean+eps); if False: score/(mean+eps)
42
+
43
+ MODEL_PATH = "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps"
44
+ DATA_PATH = "/work/nvme/bgif/gzeng/MAXRL/variance_analysis/data/SmolLM/512x512.jsonl"
45
+ MAX_SEQ_LEN = 2048 # truncate sequences longer than this
46
+ MICRO_BATCH_SIZE = 8 # for forward/backward to avoid OOM
47
+ SEED = 42
48
+ DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
49
+ DTYPE = torch.bfloat16 if torch.cuda.is_available() else torch.float32
50
+
51
+
52
+ # ============================================================================
53
+ # Data Loading
54
+ # ============================================================================
55
+ def load_rollout_data(data_path: str) -> dict:
56
+ """Load pre-computed rollouts and group by prompt.
57
+
58
+ Returns:
59
+ dict mapping prompt_id (int) -> {
60
+ "input": str,
61
+ "rollouts": [{"output": str, "score": float}, ...]
62
+ }
63
+ """
64
+ prompt_to_id = {}
65
+ prompt_data = {}
66
+
67
+ with open(data_path, "r") as f:
68
+ for line in f:
69
+ item = json.loads(line)
70
+ prompt_text = item["input"]
71
+ if prompt_text not in prompt_to_id:
72
+ pid = len(prompt_to_id)
73
+ prompt_to_id[prompt_text] = pid
74
+ prompt_data[pid] = {"input": prompt_text, "rollouts": []}
75
+ pid = prompt_to_id[prompt_text]
76
+ prompt_data[pid]["rollouts"].append({
77
+ "output": item["output"],
78
+ "score": item["score"],
79
+ })
80
+
81
+ print(f"Loaded {len(prompt_data)} prompts, "
82
+ f"each with {len(prompt_data[0]['rollouts'])} rollouts")
83
+ return prompt_data
84
+
85
+
86
+ # ============================================================================
87
+ # MaxRL Advantage Computation
88
+ # ============================================================================
89
+ def compute_maxrl_advantage(scores: list[float], epsilon: float = 1e-6) -> list[float]:
90
+ """Compute MaxRL-style advantages for a single prompt's rollouts.
91
+
92
+ This function is used to study the effect of the baseline term in MaxRL.
93
+
94
+ If BASELINE is True:
95
+ advantage_j = (score_j - mean) / (mean + epsilon)
96
+
97
+ If BASELINE is False:
98
+ advantage_j = score_j / (mean + epsilon)
99
+
100
+ In both cases, the denominator stays the same. The ablation only removes
101
+ the baseline/mean-centering term from the numerator.
102
+ """
103
+ mean = sum(scores) / len(scores)
104
+ if BASELINE:
105
+ return [(s - mean) / (mean + epsilon) for s in scores]
106
+ else:
107
+ return [(s - 0.0) / (mean + epsilon) for s in scores]
108
+
109
+
110
+ # ============================================================================
111
+ # Log Probability Computation
112
+ # ============================================================================
113
+ def tokenize_and_get_response_mask(
114
+ tokenizer,
115
+ prompt: str,
116
+ response: str,
117
+ max_seq_len: int,
118
+ ) -> tuple[torch.Tensor, torch.Tensor]:
119
+ """Tokenize prompt+response and create a response-only mask.
120
+
121
+ Returns:
122
+ input_ids: (seq_len,) token ids
123
+ response_mask: (seq_len,) binary mask, 1 for response tokens
124
+ """
125
+ prompt_ids = tokenizer.encode(prompt, add_special_tokens=False)
126
+ response_ids = tokenizer.encode(response, add_special_tokens=False)
127
+
128
+ # Truncate if needed
129
+ total_len = len(prompt_ids) + len(response_ids)
130
+ if total_len > max_seq_len:
131
+ # Keep full prompt, truncate response
132
+ max_resp = max_seq_len - len(prompt_ids)
133
+ if max_resp <= 0:
134
+ # Prompt itself is too long, truncate prompt too
135
+ prompt_ids = prompt_ids[:max_seq_len // 2]
136
+ max_resp = max_seq_len - len(prompt_ids)
137
+ response_ids = response_ids[:max_resp]
138
+
139
+ input_ids = prompt_ids + response_ids
140
+ response_mask = [0] * len(prompt_ids) + [1] * len(response_ids)
141
+
142
+ return (
143
+ torch.tensor(input_ids, dtype=torch.long),
144
+ torch.tensor(response_mask, dtype=torch.float32),
145
+ )
146
+
147
+
148
+ def pad_batch(
149
+ batch_input_ids: list[torch.Tensor],
150
+ batch_response_masks: list[torch.Tensor],
151
+ pad_token_id: int,
152
+ ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
153
+ """Pad a batch of variable-length sequences.
154
+
155
+ Returns:
156
+ input_ids: (B, max_len)
157
+ response_mask: (B, max_len)
158
+ attention_mask: (B, max_len)
159
+ """
160
+ max_len = max(ids.shape[0] for ids in batch_input_ids)
161
+ B = len(batch_input_ids)
162
+
163
+ input_ids = torch.full((B, max_len), pad_token_id, dtype=torch.long)
164
+ response_mask = torch.zeros(B, max_len)
165
+ attention_mask = torch.zeros(B, max_len)
166
+
167
+ for i, (ids, rmask) in enumerate(zip(batch_input_ids, batch_response_masks)):
168
+ seq_len = ids.shape[0]
169
+ # Left-pad: place content at the end
170
+ input_ids[i, max_len - seq_len:] = ids
171
+ response_mask[i, max_len - seq_len:] = rmask
172
+ attention_mask[i, max_len - seq_len:] = 1.0
173
+
174
+ return input_ids, response_mask, attention_mask
175
+
176
+
177
+ def compute_policy_gradient_loss(
178
+ model,
179
+ input_ids: torch.Tensor,
180
+ attention_mask: torch.Tensor,
181
+ response_mask: torch.Tensor,
182
+ advantages: torch.Tensor,
183
+ ) -> tuple[torch.Tensor, int]:
184
+ """Compute the token-mean REINFORCE loss for a micro-batch.
185
+
186
+ This matches the repo's default `loss_agg_mode=token-mean`: each response
187
+ token gets the same per-sequence advantage, and we average over valid
188
+ response tokens.
189
+
190
+ Returns:
191
+ loss: scalar loss (with grad)
192
+ valid_token_count: number of valid response tokens in this micro-batch
193
+ """
194
+ outputs = model(input_ids=input_ids, attention_mask=attention_mask)
195
+ logits = outputs.logits # (B, T, V)
196
+
197
+ # Shift: predict next token from current position
198
+ # logits[:, :-1] predicts token at position [:, 1:]
199
+ shift_logits = logits[:, :-1, :]
200
+ shift_labels = input_ids[:, 1:]
201
+ shift_response_mask = response_mask[:, 1:]
202
+
203
+ # Log probabilities of the actual tokens
204
+ log_probs = torch.log_softmax(shift_logits, dim=-1)
205
+ token_log_probs = torch.gather(
206
+ log_probs, dim=-1, index=shift_labels.unsqueeze(-1)
207
+ ).squeeze(-1) # (B, T-1)
208
+
209
+ # Token-level REINFORCE loss with a per-sequence advantage broadcast to
210
+ # every response token, then aggregated by masked token mean.
211
+ token_losses = -advantages.unsqueeze(-1) * token_log_probs * shift_response_mask
212
+ valid_token_count = int(shift_response_mask.sum().item())
213
+ loss = token_losses.sum() / max(valid_token_count, 1)
214
+ return loss, valid_token_count
215
+
216
+
217
+ # ============================================================================
218
+ # Gradient Collection Utilities
219
+ # ============================================================================
220
+ def collect_flat_gradient(model) -> torch.Tensor:
221
+ """Flatten all parameter gradients into a single vector (float32)."""
222
+ grads = []
223
+ for p in model.parameters():
224
+ if p.grad is not None:
225
+ grads.append(p.grad.detach().float().flatten())
226
+ else:
227
+ grads.append(torch.zeros(p.numel(), dtype=torch.float32, device=p.device))
228
+ return torch.cat(grads)
229
+
230
+
231
+ def compute_variance_metrics(
232
+ grad_sum: torch.Tensor,
233
+ grad_sq_sum: torch.Tensor,
234
+ K: int,
235
+ grad_norms: list[float],
236
+ grad_samples: list[torch.Tensor],
237
+ ) -> dict:
238
+ """Compute variance metrics from accumulated gradient statistics.
239
+
240
+ Args:
241
+ grad_sum: sum of K gradient vectors
242
+ grad_sq_sum: sum of K element-wise squared gradient vectors
243
+ K: number of gradient samples
244
+ grad_norms: list of gradient norms for each sample
245
+ grad_samples: list of the K gradient vectors for cosine-to-mean stats
246
+ """
247
+ grad_mean = grad_sum / K
248
+ mean_grad_norm = grad_mean.norm().item()
249
+
250
+ # Var(g) = E[g^2] - E[g]^2, then sum over dimensions -> trace of covariance
251
+ # With Bessel correction: multiply by K/(K-1)
252
+ elementwise_var = (grad_sq_sum / K - grad_mean ** 2) * (K / (K - 1))
253
+ trace_variance = elementwise_var.sum().item()
254
+
255
+ # Relative variance
256
+ relative_variance = trace_variance / (mean_grad_norm ** 2 + 1e-10)
257
+
258
+ cosine_sims_to_mean = []
259
+ if mean_grad_norm > 0:
260
+ for grad in grad_samples:
261
+ cos_sim = torch.nn.functional.cosine_similarity(
262
+ grad.unsqueeze(0), grad_mean.unsqueeze(0),
263
+ ).item()
264
+ cosine_sims_to_mean.append(cos_sim)
265
+
266
+ return {
267
+ "mean_grad_norm": mean_grad_norm,
268
+ "trace_variance": trace_variance,
269
+ "relative_variance": relative_variance,
270
+ "avg_sample_grad_norm": np.mean(grad_norms),
271
+ "std_sample_grad_norm": np.std(grad_norms),
272
+ "avg_cosine_similarity_to_mean": (
273
+ np.mean(cosine_sims_to_mean) if cosine_sims_to_mean else float("nan")
274
+ ),
275
+ }
276
+
277
+
278
+ # ============================================================================
279
+ # Main Analysis Loop
280
+ # ============================================================================
281
+ def run_variance_analysis():
282
+ random.seed(SEED)
283
+ np.random.seed(SEED)
284
+ torch.manual_seed(SEED)
285
+
286
+ # --- Load model and tokenizer ---
287
+ print(f"Loading model from {MODEL_PATH} ...")
288
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
289
+ if tokenizer.pad_token is None:
290
+ tokenizer.pad_token = tokenizer.eos_token
291
+
292
+ model = AutoModelForCausalLM.from_pretrained(
293
+ MODEL_PATH, torch_dtype=DTYPE,
294
+ ).to(DEVICE)
295
+ model.eval() # keep eval mode (no dropout), but we still need gradients
296
+ for p in model.parameters():
297
+ p.requires_grad_(True)
298
+
299
+ total_params = sum(p.numel() for p in model.parameters())
300
+ print(f"Model loaded: {total_params:,} parameters on {DEVICE}")
301
+
302
+ # --- Load data ---
303
+ print(f"Loading rollout data from {DATA_PATH} ...")
304
+ prompt_data = load_rollout_data(DATA_PATH)
305
+ all_prompt_ids = list(prompt_data.keys())
306
+
307
+ # --- Main loop ---
308
+ all_round_metrics = []
309
+
310
+ for round_idx in range(TOTAL_ROUNDS):
311
+ print(f"\n{'='*60}")
312
+ print(f"Round {round_idx + 1}/{TOTAL_ROUNDS}")
313
+ print(f"{'='*60}")
314
+
315
+ # Sample BATCH_SIZE prompts
316
+ sampled_prompts = random.sample(all_prompt_ids, BATCH_SIZE)
317
+
318
+ # Pre-sample disjoint rollout subsets for each prompt in this round.
319
+ # This keeps subset-to-subset comparisons within a round free of
320
+ # repeated rollout samples for the same prompt.
321
+ rollouts_needed_per_prompt = NUMBER_BATCHES_PER_ROUND * ROLLOUT_NUM
322
+ round_rollout_subsets = {}
323
+ for pid in sampled_prompts:
324
+ rollouts = prompt_data[pid]["rollouts"]
325
+ if len(rollouts) < rollouts_needed_per_prompt:
326
+ raise ValueError(
327
+ "Not enough rollouts for non-overlapping sampling in one round: "
328
+ f"prompt {pid} has {len(rollouts)} rollouts, but "
329
+ f"{rollouts_needed_per_prompt} are required "
330
+ f"({NUMBER_BATCHES_PER_ROUND} subsets x {ROLLOUT_NUM} rollouts)."
331
+ )
332
+
333
+ sampled_rollouts_for_round = random.sample(
334
+ rollouts, rollouts_needed_per_prompt,
335
+ )
336
+ round_rollout_subsets[pid] = [
337
+ sampled_rollouts_for_round[
338
+ subset_start:subset_start + ROLLOUT_NUM
339
+ ]
340
+ for subset_start in range(
341
+ 0, rollouts_needed_per_prompt, ROLLOUT_NUM,
342
+ )
343
+ ]
344
+
345
+ # Accumulators for this round
346
+ grad_sum = torch.zeros(total_params, dtype=torch.float32)
347
+ grad_sq_sum = torch.zeros(total_params, dtype=torch.float32)
348
+ grad_samples = []
349
+ grad_norms = []
350
+
351
+ for subset_idx in range(NUMBER_BATCHES_PER_ROUND):
352
+ print(f" Subset {subset_idx + 1}/{NUMBER_BATCHES_PER_ROUND} ...", end=" ")
353
+
354
+ # --- Sample rollouts and compute advantages ---
355
+ all_input_ids = []
356
+ all_response_masks = []
357
+ all_advantages = []
358
+
359
+ for pid in sampled_prompts:
360
+ prompt_text = prompt_data[pid]["input"]
361
+
362
+ # Use the pre-sampled non-overlapping subset for this round.
363
+ sampled_rollouts = round_rollout_subsets[pid][subset_idx]
364
+ scores = [r["score"] for r in sampled_rollouts]
365
+ advantages = compute_maxrl_advantage(scores)
366
+
367
+ for rollout, adv in zip(sampled_rollouts, advantages):
368
+ ids, rmask = tokenize_and_get_response_mask(
369
+ tokenizer, prompt_text, rollout["output"], MAX_SEQ_LEN,
370
+ )
371
+ all_input_ids.append(ids)
372
+ all_response_masks.append(rmask)
373
+ all_advantages.append(adv)
374
+
375
+ # --- Compute policy gradient via micro-batching ---
376
+ model.zero_grad()
377
+ num_samples = len(all_input_ids)
378
+ total_valid_tokens = int(
379
+ sum(rmask[1:].sum().item() for rmask in all_response_masks)
380
+ )
381
+ total_loss = 0.0
382
+
383
+ for mb_start in range(0, num_samples, MICRO_BATCH_SIZE):
384
+ mb_end = min(mb_start + MICRO_BATCH_SIZE, num_samples)
385
+
386
+ mb_ids = all_input_ids[mb_start:mb_end]
387
+ mb_masks = all_response_masks[mb_start:mb_end]
388
+ mb_advs = all_advantages[mb_start:mb_end]
389
+
390
+ input_ids, response_mask, attention_mask = pad_batch(
391
+ mb_ids, mb_masks, tokenizer.pad_token_id,
392
+ )
393
+ input_ids = input_ids.to(DEVICE)
394
+ response_mask = response_mask.to(DEVICE)
395
+ attention_mask = attention_mask.to(DEVICE)
396
+ advantages_t = torch.tensor(mb_advs, dtype=DTYPE, device=DEVICE)
397
+
398
+ mb_loss, mb_valid_tokens = compute_policy_gradient_loss(
399
+ model, input_ids, attention_mask, response_mask, advantages_t,
400
+ )
401
+ scaled_loss = mb_loss * (mb_valid_tokens / max(total_valid_tokens, 1))
402
+ scaled_loss.backward()
403
+ total_loss += mb_loss.item() * (mb_valid_tokens / max(total_valid_tokens, 1))
404
+
405
+ # --- Collect gradient ---
406
+ flat_grad = collect_flat_gradient(model).cpu()
407
+ grad_norm = flat_grad.norm().item()
408
+ grad_samples.append(flat_grad)
409
+ grad_norms.append(grad_norm)
410
+
411
+ # Update accumulators
412
+ grad_sum += flat_grad
413
+ grad_sq_sum += flat_grad ** 2
414
+
415
+ print(f"loss={total_loss:.6f}, grad_norm={grad_norm:.6f}")
416
+
417
+ # --- Compute round metrics ---
418
+ K = NUMBER_BATCHES_PER_ROUND
419
+ metrics = compute_variance_metrics(grad_sum, grad_sq_sum, K, grad_norms, grad_samples)
420
+
421
+ all_round_metrics.append(metrics)
422
+
423
+ print(f"\n Round {round_idx + 1} Results:")
424
+ print(f" Mean gradient norm: {metrics['mean_grad_norm']:.6e}")
425
+ print(f" Trace of covariance: {metrics['trace_variance']:.6e}")
426
+ print(f" Relative variance: {metrics['relative_variance']:.6e}")
427
+ print(f" Avg sample grad norm: {metrics['avg_sample_grad_norm']:.6e}")
428
+ print(f" Std sample grad norm: {metrics['std_sample_grad_norm']:.6e}")
429
+ print(
430
+ " Avg cosine sim to mean:"
431
+ f" {metrics['avg_cosine_similarity_to_mean']:.6f}"
432
+ )
433
+
434
+ # --- Average over all rounds ---
435
+ print(f"\n{'='*60}")
436
+ print(f"FINAL RESULTS (averaged over {TOTAL_ROUNDS} rounds)")
437
+ print(f"{'='*60}")
438
+ print(f" BATCH_SIZE={BATCH_SIZE}, ROLLOUT_NUM={ROLLOUT_NUM}, "
439
+ f"NUMBER_BATCHES_PER_ROUND={NUMBER_BATCHES_PER_ROUND}")
440
+
441
+ for key in all_round_metrics[0]:
442
+ values = [m[key] for m in all_round_metrics]
443
+ mean_val = np.mean(values)
444
+ std_val = np.std(values)
445
+ print(f" {key}: {mean_val:.6e} +/- {std_val:.6e}")
446
+
447
+ # --- Save results ---
448
+ output_path = os.path.join(
449
+ os.path.dirname(os.path.abspath(__file__)),
450
+ f"results_bs{BATCH_SIZE}_nr{ROLLOUT_NUM}_nb{NUMBER_BATCHES_PER_ROUND}_r{TOTAL_ROUNDS}_bl{BASELINE}.json",
451
+ )
452
+ results = {
453
+ "config": {
454
+ "batch_size": BATCH_SIZE,
455
+ "rollout_num": ROLLOUT_NUM,
456
+ "number_batches_per_round": NUMBER_BATCHES_PER_ROUND,
457
+ "total_rounds": TOTAL_ROUNDS,
458
+ "model_path": MODEL_PATH,
459
+ "max_seq_len": MAX_SEQ_LEN,
460
+ "seed": SEED,
461
+ "baseline": BASELINE,
462
+ },
463
+ "per_round": all_round_metrics,
464
+ "averaged": {
465
+ key: {
466
+ "mean": float(np.mean([m[key] for m in all_round_metrics])),
467
+ "std": float(np.std([m[key] for m in all_round_metrics])),
468
+ }
469
+ for key in all_round_metrics[0]
470
+ },
471
+ }
472
+ with open(output_path, "w") as f:
473
+ json.dump(results, f, indent=2)
474
+ print(f"\nResults saved to {output_path}")
475
+
476
+
477
+ if __name__ == "__main__":
478
+ run_variance_analysis()
smollm_variance_analysis_auto.py ADDED
@@ -0,0 +1,400 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Automated SmolLM Variance Analysis for MaxRL Policy Gradient.
3
+
4
+ Runs all 12 experiments (6 rollout_nums x 2 baseline settings) across multiple
5
+ GPUs with dynamic scheduling. Each GPU worker loads the model once and pulls
6
+ experiments from a shared task pool. Saves only trace_covariance mean/std to
7
+ outputs/ and plots the variance line chart.
8
+ """
9
+
10
+ import json
11
+ import os
12
+ import random
13
+ from functools import partial
14
+
15
+ import matplotlib
16
+ matplotlib.use("Agg")
17
+ import matplotlib.pyplot as plt
18
+ import numpy as np
19
+ import torch
20
+ import torch.multiprocessing as mp
21
+ from transformers import AutoModelForCausalLM, AutoTokenizer
22
+
23
+ # ============================================================================
24
+ # Configuration
25
+ # ============================================================================
26
+ BATCH_SIZE = 16
27
+ ROLLOUT_NUMS = [4, 8, 16, 32, 64, 128]
28
+ NUMBER_BATCHES_PER_ROUND = 4
29
+ TOTAL_ROUNDS = 32
30
+ MICRO_BATCH_SIZE = 8
31
+ MAX_SEQ_LEN = 2048
32
+ SEED = 42
33
+
34
+ MODEL_PATH = "/work/nvme/bgif/gzeng/MAXRL/checkpoints/math/smollm2_0.3B_MaxRL_gsm8k_1000_steps"
35
+ DATA_PATH = "/work/nvme/bgif/gzeng/MAXRL/variance_analysis/data/SmolLM/512x512.jsonl"
36
+
37
+ GPU_IDS = [0, 1, 2, 3]
38
+ DTYPE = torch.bfloat16
39
+
40
+ SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
41
+ OUTPUT_DIR = os.path.join(SCRIPT_DIR, "outputs", "SmolLM")
42
+
43
+
44
+ # ============================================================================
45
+ # Per-worker global state (initialized once per GPU worker)
46
+ # ============================================================================
47
+ _worker_model = None
48
+ _worker_tokenizer = None
49
+ _worker_prompt_data = None
50
+ _worker_all_prompt_ids = None
51
+ _worker_total_params = None
52
+ _worker_device = None
53
+
54
+
55
+ def worker_init(gpu_queue: mp.Queue):
56
+ """Called once per pool worker. Grabs a GPU and loads model + data."""
57
+ global _worker_model, _worker_tokenizer, _worker_prompt_data
58
+ global _worker_all_prompt_ids, _worker_total_params, _worker_device
59
+
60
+ gpu_id = gpu_queue.get()
61
+ _worker_device = f"cuda:{gpu_id}"
62
+ print(f"[Worker PID={os.getpid()}] Assigned to GPU {gpu_id}")
63
+
64
+ _worker_tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
65
+ if _worker_tokenizer.pad_token is None:
66
+ _worker_tokenizer.pad_token = _worker_tokenizer.eos_token
67
+
68
+ _worker_model = AutoModelForCausalLM.from_pretrained(
69
+ MODEL_PATH, torch_dtype=DTYPE,
70
+ ).to(_worker_device)
71
+ _worker_model.eval()
72
+ for p in _worker_model.parameters():
73
+ p.requires_grad_(True)
74
+
75
+ _worker_total_params = sum(p.numel() for p in _worker_model.parameters())
76
+ print(f"[GPU {gpu_id}] Model loaded: {_worker_total_params:,} parameters")
77
+
78
+ _worker_prompt_data = load_rollout_data(DATA_PATH)
79
+ _worker_all_prompt_ids = list(_worker_prompt_data.keys())
80
+
81
+
82
+ # ============================================================================
83
+ # Data Loading
84
+ # ============================================================================
85
+ def load_rollout_data(data_path: str) -> dict:
86
+ prompt_to_id = {}
87
+ prompt_data = {}
88
+
89
+ with open(data_path, "r") as f:
90
+ for line in f:
91
+ item = json.loads(line)
92
+ prompt_text = item["input"]
93
+ if prompt_text not in prompt_to_id:
94
+ pid = len(prompt_to_id)
95
+ prompt_to_id[prompt_text] = pid
96
+ prompt_data[pid] = {"input": prompt_text, "rollouts": []}
97
+ pid = prompt_to_id[prompt_text]
98
+ prompt_data[pid]["rollouts"].append({
99
+ "output": item["output"],
100
+ "score": item["score"],
101
+ })
102
+
103
+ print(f"Loaded {len(prompt_data)} prompts, "
104
+ f"each with {len(prompt_data[0]['rollouts'])} rollouts")
105
+ return prompt_data
106
+
107
+
108
+ # ============================================================================
109
+ # MaxRL Advantage Computation
110
+ # ============================================================================
111
+ def compute_maxrl_advantage(
112
+ scores: list[float], baseline: bool, epsilon: float = 1e-6,
113
+ ) -> list[float]:
114
+ mean = sum(scores) / len(scores)
115
+ if baseline:
116
+ return [(s - mean) / (mean + epsilon) for s in scores]
117
+ else:
118
+ return [s / (mean + epsilon) for s in scores]
119
+
120
+
121
+ # ============================================================================
122
+ # Tokenization & Batching
123
+ # ============================================================================
124
+ def tokenize_and_get_response_mask(
125
+ tokenizer, prompt: str, response: str, max_seq_len: int,
126
+ ) -> tuple[torch.Tensor, torch.Tensor]:
127
+ prompt_ids = tokenizer.encode(prompt, add_special_tokens=False)
128
+ response_ids = tokenizer.encode(response, add_special_tokens=False)
129
+
130
+ total_len = len(prompt_ids) + len(response_ids)
131
+ if total_len > max_seq_len:
132
+ max_resp = max_seq_len - len(prompt_ids)
133
+ if max_resp <= 0:
134
+ prompt_ids = prompt_ids[:max_seq_len // 2]
135
+ max_resp = max_seq_len - len(prompt_ids)
136
+ response_ids = response_ids[:max_resp]
137
+
138
+ input_ids = prompt_ids + response_ids
139
+ response_mask = [0] * len(prompt_ids) + [1] * len(response_ids)
140
+
141
+ return (
142
+ torch.tensor(input_ids, dtype=torch.long),
143
+ torch.tensor(response_mask, dtype=torch.float32),
144
+ )
145
+
146
+
147
+ def pad_batch(
148
+ batch_input_ids: list[torch.Tensor],
149
+ batch_response_masks: list[torch.Tensor],
150
+ pad_token_id: int,
151
+ ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
152
+ max_len = max(ids.shape[0] for ids in batch_input_ids)
153
+ B = len(batch_input_ids)
154
+
155
+ input_ids = torch.full((B, max_len), pad_token_id, dtype=torch.long)
156
+ response_mask = torch.zeros(B, max_len)
157
+ attention_mask = torch.zeros(B, max_len)
158
+
159
+ for i, (ids, rmask) in enumerate(zip(batch_input_ids, batch_response_masks)):
160
+ seq_len = ids.shape[0]
161
+ input_ids[i, max_len - seq_len:] = ids
162
+ response_mask[i, max_len - seq_len:] = rmask
163
+ attention_mask[i, max_len - seq_len:] = 1.0
164
+
165
+ return input_ids, response_mask, attention_mask
166
+
167
+
168
+ # ============================================================================
169
+ # Policy Gradient Loss
170
+ # ============================================================================
171
+ def compute_policy_gradient_loss(
172
+ model, input_ids, attention_mask, response_mask, advantages,
173
+ ) -> tuple[torch.Tensor, int]:
174
+ outputs = model(input_ids=input_ids, attention_mask=attention_mask)
175
+ logits = outputs.logits
176
+
177
+ shift_logits = logits[:, :-1, :]
178
+ shift_labels = input_ids[:, 1:]
179
+ shift_response_mask = response_mask[:, 1:]
180
+
181
+ log_probs = torch.log_softmax(shift_logits, dim=-1)
182
+ token_log_probs = torch.gather(
183
+ log_probs, dim=-1, index=shift_labels.unsqueeze(-1),
184
+ ).squeeze(-1)
185
+
186
+ token_losses = -advantages.unsqueeze(-1) * token_log_probs * shift_response_mask
187
+ valid_token_count = int(shift_response_mask.sum().item())
188
+ loss = token_losses.sum() / max(valid_token_count, 1)
189
+ return loss, valid_token_count
190
+
191
+
192
+ # ============================================================================
193
+ # Gradient Utilities
194
+ # ============================================================================
195
+ def collect_flat_gradient(model) -> torch.Tensor:
196
+ grads = []
197
+ for p in model.parameters():
198
+ if p.grad is not None:
199
+ grads.append(p.grad.detach().float().flatten())
200
+ else:
201
+ grads.append(torch.zeros(p.numel(), dtype=torch.float32, device=p.device))
202
+ return torch.cat(grads)
203
+
204
+
205
+ def compute_trace_variance(
206
+ grad_sum: torch.Tensor, grad_sq_sum: torch.Tensor, K: int,
207
+ ) -> float:
208
+ grad_mean = grad_sum / K
209
+ elementwise_var = (grad_sq_sum / K - grad_mean ** 2) * (K / (K - 1))
210
+ return elementwise_var.sum().item()
211
+
212
+
213
+ # ============================================================================
214
+ # Single Experiment (runs inside a worker process)
215
+ # ============================================================================
216
+ def run_single_experiment(task: tuple[int, bool]) -> tuple[str, dict]:
217
+ """Run one experiment using the worker's pre-loaded model and data.
218
+
219
+ Args:
220
+ task: (rollout_num, baseline)
221
+
222
+ Returns:
223
+ (key, {"mean": float, "std": float})
224
+ """
225
+ rollout_num, baseline = task
226
+ key = f"nr{rollout_num}_bl{baseline}"
227
+
228
+ model = _worker_model
229
+ tokenizer = _worker_tokenizer
230
+ prompt_data = _worker_prompt_data
231
+ all_prompt_ids = _worker_all_prompt_ids
232
+ total_params = _worker_total_params
233
+ device = _worker_device
234
+
235
+ print(f"[{device}] Starting {key}")
236
+
237
+ random.seed(SEED)
238
+ np.random.seed(SEED)
239
+ torch.manual_seed(SEED)
240
+
241
+ trace_variances = []
242
+
243
+ for round_idx in range(TOTAL_ROUNDS):
244
+ sampled_prompts = random.sample(all_prompt_ids, BATCH_SIZE)
245
+
246
+ rollouts_needed = NUMBER_BATCHES_PER_ROUND * rollout_num
247
+ round_rollout_subsets = {}
248
+ for pid in sampled_prompts:
249
+ rollouts = prompt_data[pid]["rollouts"]
250
+ if len(rollouts) < rollouts_needed:
251
+ raise ValueError(
252
+ f"Prompt {pid} has {len(rollouts)} rollouts, need {rollouts_needed}"
253
+ )
254
+ sampled = random.sample(rollouts, rollouts_needed)
255
+ round_rollout_subsets[pid] = [
256
+ sampled[s:s + rollout_num]
257
+ for s in range(0, rollouts_needed, rollout_num)
258
+ ]
259
+
260
+ grad_sum = torch.zeros(total_params, dtype=torch.float32)
261
+ grad_sq_sum = torch.zeros(total_params, dtype=torch.float32)
262
+
263
+ for subset_idx in range(NUMBER_BATCHES_PER_ROUND):
264
+ all_input_ids = []
265
+ all_response_masks = []
266
+ all_advantages = []
267
+
268
+ for pid in sampled_prompts:
269
+ prompt_text = prompt_data[pid]["input"]
270
+ sampled_rollouts = round_rollout_subsets[pid][subset_idx]
271
+ scores = [r["score"] for r in sampled_rollouts]
272
+ advantages = compute_maxrl_advantage(scores, baseline)
273
+
274
+ for rollout, adv in zip(sampled_rollouts, advantages):
275
+ ids, rmask = tokenize_and_get_response_mask(
276
+ tokenizer, prompt_text, rollout["output"], MAX_SEQ_LEN,
277
+ )
278
+ all_input_ids.append(ids)
279
+ all_response_masks.append(rmask)
280
+ all_advantages.append(adv)
281
+
282
+ model.zero_grad()
283
+ total_valid_tokens = int(
284
+ sum(rmask[1:].sum().item() for rmask in all_response_masks)
285
+ )
286
+ num_samples = len(all_input_ids)
287
+
288
+ for mb_start in range(0, num_samples, MICRO_BATCH_SIZE):
289
+ mb_end = min(mb_start + MICRO_BATCH_SIZE, num_samples)
290
+
291
+ mb_ids = all_input_ids[mb_start:mb_end]
292
+ mb_masks = all_response_masks[mb_start:mb_end]
293
+ mb_advs = all_advantages[mb_start:mb_end]
294
+
295
+ input_ids, response_mask, attention_mask = pad_batch(
296
+ mb_ids, mb_masks, tokenizer.pad_token_id,
297
+ )
298
+ input_ids = input_ids.to(device)
299
+ response_mask = response_mask.to(device)
300
+ attention_mask = attention_mask.to(device)
301
+ advantages_t = torch.tensor(mb_advs, dtype=DTYPE, device=device)
302
+
303
+ mb_loss, mb_valid_tokens = compute_policy_gradient_loss(
304
+ model, input_ids, attention_mask, response_mask, advantages_t,
305
+ )
306
+ scaled_loss = mb_loss * (mb_valid_tokens / max(total_valid_tokens, 1))
307
+ scaled_loss.backward()
308
+
309
+ flat_grad = collect_flat_gradient(model).cpu()
310
+ grad_sum += flat_grad
311
+ grad_sq_sum += flat_grad ** 2
312
+
313
+ trace_var = compute_trace_variance(
314
+ grad_sum, grad_sq_sum, NUMBER_BATCHES_PER_ROUND,
315
+ )
316
+ trace_variances.append(trace_var)
317
+ print(f" [{device}] {key} round {round_idx+1}/{TOTAL_ROUNDS}: "
318
+ f"trace_cov={trace_var:.6e}")
319
+
320
+ result = {
321
+ "mean": float(np.mean(trace_variances)),
322
+ "std": float(np.std(trace_variances)),
323
+ }
324
+ print(f"[{device}] Finished {key}: mean={result['mean']:.6e}, std={result['std']:.6e}")
325
+ return key, result
326
+
327
+
328
+ # ============================================================================
329
+ # Plotting
330
+ # ============================================================================
331
+ def plot_results(results: dict, output_dir: str):
332
+ rollout_nums = ROLLOUT_NUMS
333
+
334
+ means_bl_true = [results[f"nr{nr}_blTrue"]["mean"] for nr in rollout_nums]
335
+ means_bl_false = [results[f"nr{nr}_blFalse"]["mean"] for nr in rollout_nums]
336
+
337
+ fig, ax = plt.subplots(figsize=(7, 5))
338
+
339
+ ax.plot(rollout_nums, means_bl_true, marker='o', label='MaxRL')
340
+ ax.plot(rollout_nums, means_bl_false, marker='s', label='MaxRL (w/o baseline)')
341
+
342
+ ax.set_xscale('log', base=2)
343
+ ax.set_xticks(rollout_nums)
344
+ ax.set_xticklabels(rollout_nums)
345
+ ax.set_xlabel('Rollout', fontsize=14)
346
+ ax.set_ylabel('Gradient Variance', fontsize=14)
347
+ ax.legend(fontsize=12)
348
+ ax.grid(True, alpha=0.3)
349
+
350
+ plt.tight_layout()
351
+ plt.savefig(os.path.join(output_dir, "variance_plot.pdf"), dpi=300)
352
+ plt.savefig(os.path.join(output_dir, "variance_plot.png"), dpi=300)
353
+ print(f"Plots saved to {output_dir}/variance_plot.{{pdf,png}}")
354
+
355
+
356
+ # ============================================================================
357
+ # Main
358
+ # ============================================================================
359
+ def main():
360
+ os.makedirs(OUTPUT_DIR, exist_ok=True)
361
+
362
+ # Build task list: 12 experiments
363
+ tasks = []
364
+ for rollout_num in ROLLOUT_NUMS:
365
+ for baseline in [True, False]:
366
+ tasks.append((rollout_num, baseline))
367
+
368
+ print(f"Scheduling {len(tasks)} experiments across {len(GPU_IDS)} GPUs")
369
+
370
+ # GPU queue: each worker grabs one GPU ID on init
371
+ gpu_queue = mp.Queue()
372
+ for gid in GPU_IDS:
373
+ gpu_queue.put(gid)
374
+
375
+ # Pool of workers = number of GPUs. Each worker inits once (loads model),
376
+ # then processes tasks dynamically from the pool.
377
+ with mp.Pool(
378
+ processes=len(GPU_IDS),
379
+ initializer=worker_init,
380
+ initargs=(gpu_queue,),
381
+ ) as pool:
382
+ results_list = pool.map(run_single_experiment, tasks)
383
+
384
+ # Collect results
385
+ results = dict(results_list)
386
+
387
+ # Save
388
+ results_path = os.path.join(OUTPUT_DIR, "results.json")
389
+ with open(results_path, "w") as f:
390
+ json.dump(results, f, indent=2)
391
+ print(f"Results saved to {results_path}")
392
+
393
+ # Plot
394
+ plot_results(results, OUTPUT_DIR)
395
+ print("All done!")
396
+
397
+
398
+ if __name__ == "__main__":
399
+ mp.set_start_method("spawn")
400
+ main()
variance_plot.png ADDED

Git LFS Details

  • SHA256: 47d1f3ec56984b8d3f4f42e047f59d80f2988c88fb0b3993c5ab8aac5913ccb6
  • Pointer size: 131 Bytes
  • Size of remote file: 136 kB