| """ |
| EXP Q ADDENDUM: add multi-property discrete + continuous configs to the |
| scatter sweep so the headline-PosDis configs (0.76 discrete, 0.40 continuous |
| multi-prop) are represented in the metric-vs-transfer correlation. |
| |
| Trains: |
| disc_multi_L3_V5 — discrete bottleneck, 2 head per agent, V=5, multi-prop |
| (mass_bin + restit_bin, 2-headed receiver) |
| disc_multi_L4_V10 — same but L=4, V=10 (broader coverage) |
| cont_multi_dim3 — continuous bottleneck, code_dim=3 per agent, multi-prop |
| |
| Per config: |
| - Within: 3 seeds, mean across 2 heads (mass + restit) for combined acc |
| - Metrics: TopSim, PosDis, CausalSpec on multi-prop labels |
| - Cross to ramp at N=16 and N=192 (restitution only, since ramp lacks mass) |
| |
| Re-uses EXP N's MultiPropReceiver and the metric helpers in EXP N + Q. |
| """ |
| import json, time, sys, os, math |
| from pathlib import Path |
| from datetime import datetime, timezone |
| import numpy as np |
| import torch |
| import torch.nn as nn |
| import torch.nn.functional as F |
|
|
| sys.path.insert(0, os.path.dirname(__file__)) |
| from _kinematics_train import ( |
| DEVICE, ClassifierReceiver, |
| HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR, |
| EARLY_STOP_PATIENCE, |
| ) |
| from _killer_experiment import TemporalEncoder, DiscreteSender, DiscreteMultiSender |
| from _overnight_p1_transfer import make_splits |
| from _overnight_p3_matrix import load_labels, load_feat_subsampled |
| from _rev_f_cnn_control import ci95 |
| from _rev_q_posdis_scatter import build_discrete_sender, discrete_token_extract, discrete_topsim |
| from _rev_n_multiprop_continuous import ( |
| MultiPropReceiver, train_multiprop_continuous_base, |
| topsim_multiprop, posdis_multiprop, causal_spec_multiprop, |
| ) |
| from _rev_m_continuous_bottleneck import ( |
| build_continuous_sender, get_continuous_messages, |
| train_recv_frozen_cont, |
| ) |
|
|
| OUT = Path("results/reviewer_response/exp_q_addendum") |
| OUT.mkdir(parents=True, exist_ok=True) |
| N_SEEDS = 3 |
| N_LIST = [16, 192] |
|
|
|
|
| def log(msg): |
| ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ") |
| print(f"[{ts}] EXP-QADD: {msg}", flush=True) |
|
|
|
|
| |
| def train_discrete_multi(feat, labels_list, seed, n_heads, vocab_size, |
| n_epochs=150): |
| """Train DiscreteSender with multi-prop receiver (2 heads per receiver).""" |
| N, nf, dim = feat.shape |
| fpa = 1 |
| msg_dim = vocab_size * n_heads * N_AGENTS |
| agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)] |
| torch.manual_seed(seed); np.random.seed(seed) |
| rng = np.random.RandomState(seed * 1000 + 42) |
|
|
| primary = labels_list[0] |
| train_ids, holdout_ids = [], [] |
| for c in np.unique(primary): |
| ids_c = np.where(primary == c)[0] |
| rng.shuffle(ids_c) |
| split = max(1, len(ids_c) // 5) |
| holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:]) |
| train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids) |
| n_classes_per_prop = [int(lbl.max()) + 1 for lbl in labels_list] |
| chance = 1.0 / max(n_classes_per_prop) |
|
|
| sender = build_discrete_sender(dim, n_heads, vocab_size, fpa) |
| receivers = [MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE) |
| for _ in range(3)] |
| so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR) |
| ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers] |
| labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list] |
| me = math.log(vocab_size) |
| n_batches = max(1, len(train_ids) // BATCH_SIZE) |
| best_acc = 0.0; best_ep = 0 |
| best_sender_state = None; best_receiver_states = None; best_recv_idx = 0 |
|
|
| for ep in range(n_epochs): |
| if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break |
| if ep > 0 and ep % 40 == 0: |
| for i in range(len(receivers)): |
| receivers[i] = MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE) |
| ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR) |
| sender.train(); [r.train() for r in receivers] |
| tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1) |
| hard = ep >= 30 |
| rng_ep = np.random.RandomState(seed * 10000 + ep) |
| perm = rng_ep.permutation(train_ids) |
| for b in range(n_batches): |
| batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE] |
| if len(batch_ids) < 4: continue |
| views = [v[batch_ids].to(DEVICE) for v in agent_views] |
| tgts = [ld[batch_ids] for ld in labels_dev] |
| msg, logits_list = sender(views, tau=tau, hard=hard) |
| loss = torch.tensor(0.0, device=DEVICE) |
| for r in receivers: |
| head_logits = r(msg) |
| for hl, tgt in zip(head_logits, tgts): |
| loss = loss + F.cross_entropy(hl, tgt) |
| loss = loss / (len(receivers) * len(tgts)) |
| for lg in logits_list: |
| lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8) |
| ent = -(p * lp).sum(-1).mean() |
| if ent / me < 0.1: loss = loss - 0.03 * ent |
| if torch.isnan(loss): |
| so.zero_grad(); [o.zero_grad() for o in ros]; continue |
| so.zero_grad(); [o.zero_grad() for o in ros] |
| loss.backward() |
| torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0) |
| so.step(); [o.step() for o in ros] |
| if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache() |
| if (ep + 1) % 10 == 0 or ep == 0: |
| sender.eval(); [r.eval() for r in receivers] |
| with torch.no_grad(): |
| v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] |
| msg_ho, _ = sender(v_ho) |
| tgt_ho = [ld[holdout_ids] for ld in labels_dev] |
| best_per_recv = 0.0; best_idx = 0 |
| for ri, r in enumerate(receivers): |
| head_logits = r(msg_ho) |
| accs = [(hl.argmax(-1) == tgt).float().mean().item() |
| for hl, tgt in zip(head_logits, tgt_ho)] |
| combined = float(np.mean(accs)) |
| if combined > best_per_recv: |
| best_per_recv = combined; best_idx = ri |
| if best_per_recv > best_acc: |
| best_acc = best_per_recv; best_ep = ep |
| best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()} |
| best_receiver_states = [ |
| {k: v.cpu().clone() for k, v in r.state_dict().items()} |
| for r in receivers] |
| best_recv_idx = best_idx |
| return { |
| "sender_state": best_sender_state, |
| "receiver_states": best_receiver_states, |
| "best_recv_idx": best_recv_idx, |
| "train_ids": train_ids, "holdout_ids": holdout_ids, |
| "task_acc": best_acc, "chance": chance, |
| "n_classes_per_prop": n_classes_per_prop, |
| "fpa": 1, "dim": dim, |
| "n_heads": n_heads, "vocab_size": vocab_size, |
| "msg_dim": msg_dim, |
| } |
|
|
|
|
| |
| def discrete_multi_topsim(tokens, labels_list, n_pairs=5000): |
| """Spearman corr between Hamming(message tokens) and L1(label vector).""" |
| from scipy.stats import spearmanr |
| rng = np.random.RandomState(42) |
| N = tokens.shape[0] |
| n_pairs = min(n_pairs, N * (N - 1) // 2) |
| tok_d = []; lbl_d = [] |
| seen = set() |
| for _ in range(n_pairs): |
| i, j = rng.randint(0, N), rng.randint(0, N) |
| if i == j or (i, j) in seen or (j, i) in seen: continue |
| seen.add((i, j)) |
| tok_d.append(int((tokens[i] != tokens[j]).sum())) |
| lbl_d.append(sum(abs(int(lbl[i]) - int(lbl[j])) for lbl in labels_list)) |
| if len(tok_d) < 10 or np.std(tok_d) < 1e-9 or np.std(lbl_d) < 1e-9: |
| return float("nan") |
| rho, _ = spearmanr(tok_d, lbl_d) |
| return float(rho) if not np.isnan(rho) else 0.0 |
|
|
|
|
| def _mi_disc(x, y): |
| n = len(x) |
| n_x = int(np.max(x)) + 1; n_y = int(np.max(y)) + 1 |
| p_x = np.bincount(x, minlength=n_x) / n |
| p_y = np.bincount(y, minlength=n_y) / n |
| H_x = -np.sum([p * np.log(p) for p in p_x if p > 0]) |
| H_y = -np.sum([p * np.log(p) for p in p_y if p > 0]) |
| joint = np.zeros((n_x, n_y)) |
| for xv, yv in zip(x, y): joint[int(xv), int(yv)] += 1 |
| joint /= n |
| H_xy = 0.0 |
| for v in joint.ravel(): |
| if v > 0: H_xy -= v * np.log(v) |
| return max(H_x + H_y - H_xy, 0.0) |
|
|
|
|
| def discrete_multi_posdis(tokens, labels_list): |
| """Standard PosDis on discrete tokens: per-position, MI with each |
| property; PosDis = mean over positions of (top-second)/top.""" |
| P = tokens.shape[1] |
| K = len(labels_list) |
| mi_matrix = np.zeros((P, K)) |
| for p in range(P): |
| for k in range(K): |
| mi_matrix[p, k] = _mi_disc(tokens[:, p], labels_list[k]) |
| if mi_matrix.sum() < 1e-9: return float("nan"), mi_matrix |
| n_active = 0; total = 0.0 |
| for p in range(P): |
| sorted_mi = np.sort(mi_matrix[p])[::-1] |
| if sorted_mi[0] > 1e-6: |
| total += (sorted_mi[0] - sorted_mi[1]) / sorted_mi[0] |
| n_active += 1 |
| if n_active == 0: return float("nan"), mi_matrix |
| return float(total / n_active), mi_matrix |
|
|
|
|
| def discrete_multi_causal(base, feat, labels_list, holdout_ids): |
| """Mask each (agent x head) block; per-property accuracy drop.""" |
| sender = build_discrete_sender(feat.shape[2], base["n_heads"], |
| base["vocab_size"], base["fpa"]) |
| sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE) |
| receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM, |
| base["n_classes_per_prop"]).to(DEVICE) |
| for _ in range(len(base["receiver_states"]))] |
| for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s) |
| [r.eval() for r in receivers] |
| best_recv = receivers[base.get("best_recv_idx", 0)] |
| agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)] |
| labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list] |
| K = len(labels_list); V = base["vocab_size"]; H = base["n_heads"] |
| n_positions = N_AGENTS * H |
| with torch.no_grad(): |
| v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] |
| msg_ho, _ = sender(v_ho) |
| tgt_ho = [ld[holdout_ids] for ld in labels_dev] |
| baseline_per_prop = [(hl.argmax(-1) == tgt).float().mean().item() |
| for hl, tgt in zip(best_recv(msg_ho), tgt_ho)] |
| drops = np.zeros((n_positions, K)) |
| for pos in range(n_positions): |
| masked = msg_ho.clone() |
| start = pos * V; end = start + V |
| mean_block = msg_ho[:, start:end].mean(dim=0) |
| masked[:, start:end] = mean_block |
| for p_idx, (hl, tgt) in enumerate(zip(best_recv(masked), tgt_ho)): |
| acc = (hl.argmax(-1) == tgt).float().mean().item() |
| drops[pos, p_idx] = baseline_per_prop[p_idx] - acc |
| return baseline_per_prop, drops |
|
|
|
|
| |
| def disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, ho_ids, restit_idx=1): |
| """Apply discrete-multi sender + best receiver's restit head to target.""" |
| sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"], |
| base["vocab_size"], base["fpa"]) |
| sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE) |
| receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM, |
| base["n_classes_per_prop"]).to(DEVICE) |
| for _ in range(len(base["receiver_states"]))] |
| for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s) |
| [r.eval() for r in receivers] |
| agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)] |
| labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE) |
| with torch.no_grad(): |
| v_ho = [v[ho_ids].to(DEVICE) for v in agent_views] |
| msg_ho, _ = sender(v_ho) |
| tgt_ho = labels_dev[ho_ids] |
| best = 0.0 |
| for r in receivers: |
| head_logits = r(msg_ho) |
| preds = head_logits[restit_idx].argmax(-1) |
| acc = (preds == tgt_ho).float().mean().item() |
| if acc > best: best = acc |
| return best |
|
|
|
|
| def disc_multi_train_recv_frozen(base, feat_tgt, labels_tgt, train_ids, holdout_ids, |
| seed, n_target, n_epochs=80): |
| """Freeze sender; train fresh single-property receiver on n_target stratified |
| target examples; eval on holdout. Mirrors disc_train_recv_custom but uses |
| multi-prop sender.""" |
| if n_target == 0: |
| return disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, holdout_ids) |
| rng = np.random.RandomState(seed * 311 + 7 + n_target) |
| n_t_classes = int(np.max(labels_tgt)) + 1 |
| per_class = max(1, n_target // n_t_classes) |
| picks = [] |
| for c in range(n_t_classes): |
| ids_c = np.array([i for i in train_ids if labels_tgt[i] == c]) |
| if len(ids_c) == 0: continue |
| rng.shuffle(ids_c) |
| picks.extend(ids_c[:per_class]) |
| picks = np.array(picks) |
| if len(picks) > n_target: picks = picks[:n_target] |
| elif len(picks) < n_target and len(train_ids) > len(picks): |
| extras = np.array([i for i in train_ids if i not in set(picks)]) |
| rng.shuffle(extras) |
| picks = np.concatenate([picks, extras[:n_target - len(picks)]]) |
| if len(picks) < 2: return float("nan") |
| sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"], |
| base["vocab_size"], base["fpa"]) |
| sender.load_state_dict(base["sender_state"]); sender.to(DEVICE).eval() |
| for p in sender.parameters(): p.requires_grad = False |
| receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, n_t_classes).to(DEVICE) |
| for _ in range(3)] |
| ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers] |
| agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)] |
| labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE) |
| bs = min(BATCH_SIZE, len(picks)) |
| best = 0.0 |
| for ep in range(n_epochs): |
| [r.train() for r in receivers] |
| rng_ep = np.random.RandomState(seed * 10000 + ep) |
| perm = rng_ep.permutation(picks) |
| for b in range(max(1, len(picks) // bs)): |
| batch = perm[b*bs:(b+1)*bs] |
| if len(batch) < 2: continue |
| views = [v[batch].to(DEVICE) for v in agent_views] |
| with torch.no_grad(): |
| msg, _ = sender(views) |
| for r, o in zip(receivers, ros): |
| logits = r(msg) |
| loss = F.cross_entropy(logits, labels_dev[batch]) |
| if torch.isnan(loss): continue |
| o.zero_grad(); loss.backward(); o.step() |
| if (ep + 1) % 5 == 0: |
| [r.eval() for r in receivers] |
| with torch.no_grad(): |
| v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] |
| msg_ho, _ = sender(v_ho) |
| tgt_ho = labels_dev[holdout_ids] |
| for r in receivers: |
| preds = r(msg_ho).argmax(-1) |
| acc = (preds == tgt_ho).float().mean().item() |
| if acc > best: best = acc |
| return best |
|
|
|
|
| |
| def main(): |
| t0 = time.time() |
| log("=" * 60) |
| log("EXP Q ADDENDUM: multi-property bottleneck configs") |
|
|
| feat_c = load_feat_subsampled("collision", "vjepa2") |
| feat_r = load_feat_subsampled("ramp", "vjepa2") |
| lbl_c_mass = load_labels("collision", "mass") |
| lbl_c_rest = load_labels("collision", "restitution") |
| lbl_r_rest = load_labels("ramp", "restitution") |
| log(f" collision: feat={tuple(feat_c.shape)} mass={np.bincount(lbl_c_mass).tolist()} " |
| f"rest={np.bincount(lbl_c_rest).tolist()}") |
|
|
| rows = [] |
|
|
| |
| discrete_specs = [ |
| ("disc_multi_L3_V5", 3, 5), |
| ("disc_multi_L4_V10", 4, 10), |
| ] |
| for name, H, V in discrete_specs: |
| log(f"\n --- {name} (L={H}, V={V}, multi-prop) ---") |
| within_accs = []; bases = [] |
| for seed in range(N_SEEDS): |
| t_s = time.time() |
| try: |
| base = train_discrete_multi(feat_c, [lbl_c_mass, lbl_c_rest], |
| seed, H, V) |
| bases.append(base); within_accs.append(float(base["task_acc"])) |
| log(f" {name} s{seed}: combined within={base['task_acc']:.3f} " |
| f"[{time.time()-t_s:.0f}s]") |
| except Exception as e: |
| log(f" {name} s{seed} FAILED: {e}") |
| bases.append(None); within_accs.append(float("nan")) |
| valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)] |
| if not valid: |
| log(f" {name}: no successful base"); continue |
| best_idx = max(valid, key=lambda x: x[1])[0] |
| best_base = bases[best_idx] |
| ho_ids = best_base["holdout_ids"] |
| |
| try: |
| tokens = discrete_token_extract(best_base, feat_c) |
| tokens_ho = tokens[ho_ids] |
| ts = discrete_multi_topsim(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) |
| pd_, mi = discrete_multi_posdis(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) |
| base_pp, drops = discrete_multi_causal(best_base, feat_c, |
| [lbl_c_mass, lbl_c_rest], ho_ids) |
| cs = float(drops.max()) |
| except Exception as e: |
| log(f" {name} metrics FAILED: {e}") |
| ts = pd_ = cs = float("nan") |
| |
| cross = {n: [] for n in N_LIST} |
| for seed, base in enumerate(bases): |
| if base is None: |
| for n in N_LIST: cross[n].append(float("nan")) |
| continue |
| tr_t, ho_t = make_splits(lbl_r_rest, seed) |
| for n in N_LIST: |
| try: |
| acc = disc_multi_train_recv_frozen(base, feat_r, lbl_r_rest, |
| tr_t, ho_t, seed, n) |
| cross[n].append(float(acc)) |
| except Exception as e: |
| log(f" {name} s{seed} N={n} FAILED: {e}") |
| cross[n].append(float("nan")) |
| wm = float(np.mean([a for a in within_accs if not np.isnan(a)])) |
| cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)])) |
| if any(not np.isnan(x) for x in cross[n]) else float("nan") |
| for n in N_LIST} |
| log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} " |
| f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}") |
| rows.append({ |
| "name": name, "type": "discrete_multi", |
| "n_heads": H, "vocab_size": V, |
| "msg_dim": V * H * N_AGENTS, |
| "within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs, |
| "cross_n16": cm[16], "cross_n192": cm[192], |
| }) |
|
|
| |
| cont_spec = ("cont_multi_dim3", 3) |
| name, D = cont_spec |
| log(f"\n --- {name} (D={D}, multi-prop) ---") |
| within_accs = []; bases = [] |
| for seed in range(N_SEEDS): |
| t_s = time.time() |
| try: |
| base = train_multiprop_continuous_base( |
| feat_c, [lbl_c_mass, lbl_c_rest], seed, |
| code_dim_per_agent=D, n_epochs=150) |
| bases.append(base); within_accs.append(float(base["task_acc"])) |
| log(f" {name} s{seed}: combined within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]") |
| except Exception as e: |
| log(f" {name} s{seed} FAILED: {e}") |
| bases.append(None); within_accs.append(float("nan")) |
| valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)] |
| if valid: |
| best_idx = max(valid, key=lambda x: x[1])[0] |
| best_base = bases[best_idx] |
| ho_ids = best_base["holdout_ids"] |
| try: |
| msgs = get_continuous_messages(best_base, feat_c) |
| msgs_ho = msgs[ho_ids] |
| ts = topsim_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) |
| pd_, _ = posdis_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) |
| base_pp, drops = causal_spec_multiprop(best_base, feat_c, |
| [lbl_c_mass, lbl_c_rest], ho_ids) |
| cs = float(drops.max()) |
| except Exception as e: |
| log(f" {name} metrics FAILED: {e}") |
| ts = pd_ = cs = float("nan") |
| |
| cross = {n: [] for n in N_LIST} |
| for seed, base in enumerate(bases): |
| if base is None: |
| for n in N_LIST: cross[n].append(float("nan")) |
| continue |
| single_base = dict(base) |
| single_base["n_classes"] = base["n_classes_per_prop"][1] |
| single_base["receiver_states"] = [] |
| tr_t, ho_t = make_splits(lbl_r_rest, seed) |
| for n in N_LIST: |
| try: |
| acc = train_recv_frozen_cont(single_base, feat_r, lbl_r_rest, |
| tr_t, ho_t, seed, n) |
| cross[n].append(float(acc)) |
| except Exception as e: |
| log(f" {name} s{seed} N={n} FAILED: {e}") |
| cross[n].append(float("nan")) |
| wm = float(np.mean([a for a in within_accs if not np.isnan(a)])) |
| cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)])) |
| if any(not np.isnan(x) for x in cross[n]) else float("nan") |
| for n in N_LIST} |
| log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} " |
| f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}") |
| rows.append({ |
| "name": name, "type": "continuous_multi", |
| "code_dim": D, "msg_dim": D * N_AGENTS, |
| "within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs, |
| "cross_n16": cm[16], "cross_n192": cm[192], |
| }) |
|
|
| |
| original_q_rows = [ |
| |
| ("disc_L2_V5", "discrete", 0.88, 0.20, 0.02, 41.7, 43.9), |
| ("disc_L2_V10", "discrete", 0.84, 0.25, 0.05, 46.1, 41.7), |
| ("disc_L3_V5", "discrete", 0.84, 0.13, 0.02, 43.3, 42.8), |
| ("disc_L3_V10", "discrete", 0.84, 0.12, 0.01, 43.3, 45.6), |
| ("disc_L4_V5", "discrete", 0.90, 0.10, 0.01, 41.1, 42.2), |
| ("disc_L4_V10", "discrete", 0.82, 0.08, 0.02, 45.0, 45.0), |
| ("disc_L5_V5", "discrete", 0.89, 0.07, 0.02, 40.0, 43.9), |
| ("cont_dim2", "continuous", 0.92, 0.15, 0.20, 48.9, 54.4), |
| ("cont_dim3", "continuous", 0.91, 0.15, 0.02, 40.6, 41.1), |
| ("cont_dim5", "continuous", 0.89, 0.06, 0.03, 47.2, 43.9), |
| ("cont_dim10", "continuous", 0.88, 0.04, 0.01, 47.8, 48.3), |
| ("cont_dim20", "continuous", 0.90, 0.02, 0.00, 48.9, 55.0), |
| ] |
| all_rows = list(original_q_rows) |
| for r in rows: |
| all_rows.append(( |
| r["name"], r["type"], r["topsim"], r["posdis"], r["causal_spec"], |
| r["cross_n16"] * 100 if r["cross_n16"] <= 1 else r["cross_n16"], |
| r["cross_n192"] * 100 if r["cross_n192"] <= 1 else r["cross_n192"], |
| )) |
|
|
| |
| from scipy.stats import spearmanr |
| def safe_corr(idx_x, idx_y): |
| x = []; y = [] |
| for r in all_rows: |
| if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y])): |
| x.append(r[idx_x]); y.append(r[idx_y]) |
| if len(x) < 4 or np.std(x) < 1e-9 or np.std(y) < 1e-9: |
| return float("nan"), float("nan") |
| rho, p = spearmanr(x, y) |
| return float(rho), float(p) |
| corrs = { |
| ("topsim", 16): safe_corr(2, 5), |
| ("topsim", 192): safe_corr(2, 6), |
| ("posdis", 16): safe_corr(3, 5), |
| ("posdis", 192): safe_corr(3, 6), |
| ("causal", 16): safe_corr(4, 5), |
| ("causal", 192): safe_corr(4, 6), |
| } |
|
|
| |
| lines = [ |
| "EXP Q ADDENDUM -- multi-property bottleneck configs", |
| "", |
| f"{'Config':<22s} | {'TopSim':<8s} | {'PosDis':<8s} | {'CausalSpec':<12s} | " |
| f"{'Cross 16':<10s} | {'Cross 192':<10s}", |
| "-" * 90, |
| ] |
| for r in all_rows: |
| name, typ, ts, pd_, cs, c16, c192 = r |
| lines.append(f"{name:<22s} | {ts:+.2f} | {pd_:.2f} | {cs:.2f} " |
| f"| {c16:5.1f}% | {c192:5.1f}%") |
|
|
| lines.append("") |
| lines.append("FULL-SWEEP SPEARMAN (now including multi-prop configs):") |
| for tgt_n in [16, 192]: |
| lines.append(f" vs cross_n{tgt_n}:") |
| for met, label in [("topsim", "TopSim"), ("posdis", "PosDis"), |
| ("causal", "CausalSpec")]: |
| rho, p = corrs[(met, tgt_n)] |
| lines.append(f" {label:<12s}: rho={rho:+.2f} p={p:.3f}") |
|
|
| abs_max_rho = 0 |
| for k, (rho, p) in corrs.items(): |
| if not np.isnan(rho): abs_max_rho = max(abs_max_rho, abs(rho)) |
| lines.append("") |
| if abs_max_rho < 0.30: |
| verd = "Metrics still uncorrelated with transfer." |
| elif abs_max_rho < 0.55: |
| verd = f"Weak/moderate correlation (max |rho|={abs_max_rho:.2f})." |
| else: |
| verd = f"Strong correlation (max |rho|={abs_max_rho:.2f}). Reframe." |
| lines.append(f"VERDICT: {verd}") |
| lines.append(f"\nTotal runtime: {(time.time()-t0)/60:.1f} min") |
|
|
| summary = "\n".join(lines) |
| (OUT / "exp_q_addendum_summary.txt").write_text(summary + "\n") |
| (OUT / "exp_q_addendum_summary.json").write_text(json.dumps({ |
| "new_rows": rows, |
| "all_rows_combined": [{"name": r[0], "type": r[1], "topsim": r[2], |
| "posdis": r[3], "causal_spec": r[4], |
| "cross_n16": r[5], "cross_n192": r[6]} |
| for r in all_rows], |
| "spearman": {f"{m}__n{n}": list(v) for (m, n), v in corrs.items()}, |
| }, indent=2, default=str)) |
| print("\n" + summary, flush=True) |
| log(f"DONE in {(time.time()-t0)/60:.1f} min") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|