benchmark / docs /PHASE5_OPS.md
kengzwl's picture
Upload folder using huggingface_hub
e313100 verified

Phase 5.1 PPO — Operations Tracker

Single source of truth for in-flight work. Resume from here.


Principles

  1. Two canonical specs: ppo_playground (DM Control) and ppo_playground_loco (Loco). Per-env variants only when officially required: ppo_playground_fingerspin (gamma=0.95), ppo_playground_pendulum (training_epoch=4, action_repeat=4 via code).
  2. 100M frames hard cap — no extended runs. If an env doesn't hit target at 100M, fix the spec.
  3. Strategic reruns: only rerun failing/⚠️ envs. Already-✅ envs skip revalidation.
  4. Score metric: use total_reward_ma (final moving average of total reward) — measures end-of-training performance and matches mujoco_playground reference scores.
  5. Official reference: check ~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py for per-env overrides.

Wave 3 (2026-03-16)

Fixes applied:

  • stderr suppression: MuJoCo C-level warnings (ccd_iterations, nefc overflow, broadphase overflow) silenced in playground.py
  • obs fix: _get_obs now passes only "state" key for dict-obs envs (was incorrectly concatenating privileged_state+state)

Envs graduated to ✅ (close enough): FishSwim, PointMass, ReacherHard, WalkerStand, WalkerWalk, SpotGetup, SpotJoystickGaitTracking, AlohaHandOver

Failing envs by root cause:

  • Humanoid double-norm (rs10 fix): HumanoidStand (114→700), HumanoidWalk (47→500), HumanoidRun (18→130)
  • Dict obs fix (now fixed): Go1Flat/Rough/Getup/Handstand, G1Flat/Rough, T1Flat/Rough
  • Unknown: BarkourJoystick (0/35), Op3Joystick (0/20)
  • Needs hparam work: H1Inplace (4→10), H1Joystick (16→30), SpotFlat (11→30)
  • Manipulation: AlohaPeg (188→300), LeapCubeReorient (74→200)
  • Infeasible: PandaRobotiqPushCube, AeroCubeRotateZAxis

Currently running: (to be populated by ops)


Currently Running (as of 2026-03-14 ~00:00)

Wave V (p5-ppo17) — Constant LR test (4 runs, just launched)

Testing constant LR (Brax default) in isolation — never tested before. Key hypothesis: LR decay hurts late-converging envs.

Run Env Spec Key Change Old Best Target
p5-ppo17-csup CartpoleSwingup constlr constant LR + minibatch=4096 576.1 800
p5-ppo17-csupsparse CartpoleSwingupSparse constlr constant LR + minibatch=4096 296.3 425
p5-ppo17-acrobot AcrobotSwingup vnorm_constlr constant LR + vnorm 173 220
p5-ppo17-fteasy FingerTurnEasy vnorm_constlr constant LR + vnorm 571 950

Wave IV-H (p5-ppo16h) — Humanoid with wider policy (3 runs, ~2.5h remaining)

New ppo_playground_humanoid variant: 2×256 policy (vs 2×64), constant LR, vnorm=true. Based on Phase 3 Gymnasium Humanoid-v5 success (2661 MA with 2×256 + constant LR).

Run Env Old Best Target
p5-ppo16h-hstand HumanoidStand 18.36 700
p5-ppo16h-hwalk HumanoidWalk 7.68 500
p5-ppo16h-hrun HumanoidRun 3.19 130

Wave VI (p5-ppo18) — Brax 4×32 policy + constant LR + vnorm (3 runs, just launched)

Testing Brax default policy architecture (4 layers × 32 units vs our 2 × 64). Deeper narrower policy may learn better features for precision tasks.

Run Env Old Best Target
p5-ppo18-fteasy FingerTurnEasy 571 950
p5-ppo18-fthard FingerTurnHard 484 950
p5-ppo18-fishswim FishSwim 463 650

Wave IV tail (p5-ppo16) — completed

Run Env strength Target New best?
p5-ppo16-swimmer6 SwimmerSwimmer6 509.3 560 ✅ New best (final_strength=560.6)
p5-ppo16-fishswim FishSwim 420.6 650 ❌ Worse than 463

Wave IV results (p5-ppo16, vnorm=true rerun with reverted spec — completed):

All ran with vnorm=true. CartpoleSwingup/Sparse worse (vnorm=false is better for them — wrong setting). Precision envs also scored below old bests. Humanoid still failing with standard 2×64 policy.

Env p16 strength Old Best Target Verdict
CartpoleSwingup 316.2 576.1 (false) 800 ❌ wrong vnorm
CartpoleSwingupSparse 288.7 296.3 (false) 425 ❌ wrong vnorm
AcrobotSwingup 145.4 173 (true) 220 ❌ worse
FingerTurnEasy 511.1 571 (true) 950 ❌ worse
FingerTurnHard 368.6 484 (true) 950 ❌ worse
HumanoidStand 12.72 18.36 700 ❌ still failing
HumanoidWalk 7.46 7.68 500 ❌ still failing
HumanoidRun 3.19 3.19 130 ❌ still failing

CONCLUSION: Reverted spec didn't help. No new bests. Consistency was negative for CartpoleSwingup/Sparse (high variance). Need constant LR test (Wave V) and wider policy for Humanoid (Wave IV-H).

Wave III results (p5-ppo13/p5-ppo15, 5-layer value + no grad clip — completed):

Only CartpoleSwingup improved slightly (623.8 vs 576.1). All others regressed. FishSwim p5-ppo15: strength=411.6 (vs 463 old best). AcrobotSwingup p5-ppo15: strength=95.4 (vs 173).

CONCLUSION: 5-layer value + no grad clip is NOT a general improvement. Reverted to 3-layer + clip_grad_val=1.0.

Wave H results (p5-ppo12, ALL completed — NONE improved over old bests): Re-running same spec (variance reruns + vnorm) didn't help. Run-to-run variance is high but old bests represent lucky runs. Hyperparameter tuning has hit diminishing returns.

Wave G/G2 results (normalize_v_targets=false ablation, ALL completed):

Env p11 strength Old Best (true) Target Change Verdict
PendulumSwingup 533.5 276 395 +93% ✅ NEW PASS
FingerSpin 652.4 561 600 +16% ✅ NEW PASS
CartpoleBalanceSparse 690.4 545 700 +27% ⚠️ 99% of target
CartpoleSwingup 576.1 443/506 800 +30% ⚠️ improved
CartpoleSwingupSparse 296.3 271 425 +9% ⚠️ improved
PointMass 854.4 863 900 -1% ⚠️ same
FishSwim 293.9 463 650 -36% ❌ regression
FingerTurnEasy 441.1 571 950 -23% ❌ regression
SwimmerSwimmer6 386.2 485 560 -20% ❌ regression
FingerTurnHard 335.7 484 950 -31% ❌ regression
AcrobotSwingup 105.1 173 220 -39% ❌ regression
HumanoidStand 12.87 18.36 500 -30% ❌ still failing

CONCLUSION: normalize_v_targets: false helps 5/12, hurts 6/12, neutral 1/12.

  • false wins: PendulumSwingup, FingerSpin, CartpoleBalanceSparse, CartpoleSwingup, CartpoleSwingupSparse
  • true wins: FishSwim, FingerTurnEasy/Hard, SwimmerSwimmer6, AcrobotSwingup, PointMass
  • Decision: Per-env spec selection. New ppo_playground_vnorm variant for precision envs.

Wave F results (multi-unroll=16 + proven hyperparameters):

Env p10 strength p10 final_str Old best str Target Verdict
CartpoleSwingup 342 443 443 800 Same
FingerTurnEasy 529 685 571 950 Better final, worse strength
FingerSpin 402 597 561 600 Better final (near target!), worse strength
FingerTurnHard 368 559 484 950 Better final, worse strength
SwimmerSwimmer6 251 384 485 560 Worse
CartpoleSwingupSparse 56 158 271 425 MUCH worse
AcrobotSwingup 31 63 173 220 MUCH worse

CONCLUSION: Multi-unroll adds no benefit over single-unroll for any env by strength metric. The final_strength improvements for Finger tasks are offset by strength regressions. Root cause: stale old_net (480 vs 30 steps between copies) makes policy ratio less accurate. Spec reverted to single-unroll (num_unrolls=1). Multi-unroll code preserved in ppo.py.

Wave E results (multi-unroll + Brax hyperparameters — ALL worse):

Brax-matched spec (clip_eps=0.3, constant LR, 5-layer value, reward_scale=10, minibatch=30720) hurt every env except HopperStand (which used wrong spec before). Reverted.

Wave C completed results (all reward_scale=10, divide by 10 for true score):

Run Env strength/10 final_strength/10 total_reward_ma/10 Target vs Old
p5-ppo7-cartpoleswingup CartpoleSwingup 556.6 670.5 705.3 800 443→557 ✅ improved
p5-ppo7-fingerturneasy FingerTurnEasy 511.1 693.2 687.0 950 571→511 ❌ WORSE
p5-ppo7-fingerturnhard FingerTurnHard 321.9 416.8 425.2 950 484→322 ❌ WORSE
p5-ppo7-cartpoleswingupsparse2 CartpoleSwingupSparse 144.0 360.6 337.7 425 271→144 ❌ WORSE

KEY FINDING: time_horizon=480 helps CartpoleSwingup (+25%) but HURTS FingerTurn (-30 to -50%) and CartpoleSwingupSparse (-47%). Long GAE horizons produce noisy advantage estimates for precision/sparse tasks. The official Brax approach is 16×30-step unrolls (short GAE per unroll), NOT 1×480-step unroll.


Spec Changes Applied (2026-03-13)

Fix 1: reward_scale=10.0 (matches official mujoco_playground)

  • playground.py: PlaygroundVecEnv now multiplies rewards by self._reward_scale
  • __init__.py: threads reward_scale from env spec to wrapper
  • ppo_playground.yaml: reward_scale: 10.0 in shared _env anchor

Fix 2: Revert minibatch_size 2048→4096 (fixes CartpoleSwingup regression)

  • ppo_playground.yaml: all DM Control specs (ppo_playground, fingerspin, pendulum) now use minibatch_size=4096
  • 15 minibatches × 16 epochs = 240 grad steps (was 30×16=480)
  • Restores p5-ppo5 performance for CartpoleSwingup (803 vs 443)

Fix 3: Brax-matched spec (commit 6eb08fe9) — time_horizon=480, clip_eps=0.3, constant LR, 5-layer value net

  • Increased time_horizon from 30→480 to match total data per update (983K transitions)
  • clip_eps 0.2→0.3, constant LR (min_factor=1.0), 5-layer [256×5] value net
  • action std upper bound raised (max=2.0 in policy_util.py)
  • Result: CartpoleSwingup improved (443→557 strength), but FingerTurn and CartpoleSwingupSparse got WORSE
  • Root cause: 1×480-step unroll computes GAE over 480 steps (noisy), vs official 16×30-step unrolls (short, accurate GAE)

Fix 4: ppo_playground_short variant (time_horizon=30 + Brax improvements)

  • Keeps: reward_scale=10, clip_eps=0.3, constant LR, 5-layer value net, no grad clipping
  • Reverts: time_horizon=30, minibatch_size=4096 (15 minibatches, 240 grad steps)
  • Hypothesis: Short GAE + other Brax improvements = best of both worlds for precision tasks
  • Testing on FingerTurnEasy/Hard first (Wave D p5-ppo8-*)

Fix 5: Multi-unroll collection (IMPLEMENTED but NOT USED — code stays, spec reverted)

  • Added num_unrolls parameter to PPO (ppo.py, actor_critic.py). Code works correctly.
  • Brax-matched spec (Wave E, p5-ppo9): clip_eps=0.3, constant LR, 5-layer value, reward_scale=10
    • Result: WORSE on 5/7 tested envs. Only CartpoleSwingup improved (443→506).
    • Root cause: minibatch_size=30720 → 7.5x fewer gradient steps per transition → underfitting
  • Reverted spec + multi-unroll (Wave F, p5-ppo10): clip_eps=0.2, LR decay, 3-layer value, minibatch=4096
    • Result: Same or WORSE on all envs by strength metric. Same fps as single-unroll.
    • Training compute per env step is identical, but old_net staleness (480 vs 30 steps) hurts.
  • Conclusion: Multi-unroll adds complexity without benefit. Reverted spec to single-unroll (num_unrolls=1). Code preserved in ppo.py (defaults to 1). Spec uses original hyperparameters.

Completed Runs Needing Intake

Humanoid (ppo_playground_loco, post log_std fix) — intake immediately

Run HF Folder strength target HF status
p5-ppo6-humanoidrun ppo_playground_loco_humanoidrun_2026_03_12_175917 2.78 130 ✅ uploaded
p5-ppo6-humanoidwalk ppo_playground_loco_humanoidwalk_2026_03_12_175817 6.82 500 ✅ uploaded
p5-ppo6-humanoidstand ppo_playground_loco_humanoidstand_2026_03_12_175810 12.45 700 UPLOAD FAILED (412) — re-upload first

Re-upload HumanoidStand:

source .env && huggingface-cli upload SLM-Lab/benchmark-dev \
  hf_data/data/benchmark-dev/data/ppo_playground_loco_humanoidstand_2026_03_12_175810 \
  data/ppo_playground_loco_humanoidstand_2026_03_12_175810 --repo-type dataset

Conclusion: loco spec still fails completely for Humanoid — log_std fix insufficient. See spec fixes below.

BENCHMARKS.md correction needed (commit b6ef49d9 used wrong metric)

intake-a used total_reward_ma instead of strength. Fix these 4 entries:

Env Run strength (correct) total_reward_ma (wrong) target
AcrobotSwingup p5-ppo6-acrobotswingup2 172.8 253.24 220
CartpoleBalanceSparse p5-ppo6-cartpolebalancesparse2 545.1 991.81 700
CartpoleSwingup p5-ppo6-cartpoleswingup2 unknown — extract from logs 641.51 800
CartpoleSwingupSparse p5-ppo6-cartpoleswingupsparse 270.9 331.23 425

Extract correct values: dstack logs p5-ppo6-NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1 → use strength: field.

Also check FingerSpin: dstack logs p5-ppo6-fingerspin2 --since 6h | grep trial_metrics | tail -1 — confirm strength value.

Metric decision needed: strength penalizes slow learners (CartpoleBalanceSparse strength=545 but final MA=992). Consider switching ALL entries to final_strength. But this requires auditing every existing entry — do it as a batch before publishing.


Queue (launch when slots open, all 100M)

Priority Env Spec Run name Rationale
1 PendulumSwingup ppo_playground_pendulum p5-ppo6-pendulumswingup action_repeat=4 + training_epoch=4 (code fix applied)
2 FingerSpin ppo_playground_fingerspin p5-ppo6-fingerspin3 canonical gamma=0.95 run; fingerspin2 used gamma=0.995 (override silently ignored)

Full Env Status

✅ Complete (13/25)

Env strength target normalize_v_targets
CartpoleBalance 968.23 950 true
AcrobotSwingupSparse 42.74 15 true
BallInCup 942.44 680 true
CheetahRun 865.83 850 true
ReacherEasy 955.08 950 true
ReacherHard 946.99 950 true
WalkerRun 637.80 560 true
WalkerStand 970.94 1000 true
WalkerWalk 952 960 true
HopperHop 22.00 ~2 true
HopperStand 118.2 ~70 true
PendulumSwingup 533.5 395 false
FingerSpin 652.4 600 false

⚠️ Below target (9/25)

Env best strength target best with status
CartpoleSwingup 576.1 800 false Improved +30% from 443 (true)
CartpoleBalanceSparse 545 700 true Testing false (p5-ppo11)
CartpoleSwingupSparse 296.3 425 false Improved +9% from 271 (true)
AcrobotSwingup 173 220 true false=105, regressed
FingerTurnEasy 571 950 true false=441, regressed
FingerTurnHard 484 950 true false=336, regressed
FishSwim 463 650 true Testing false (p5-ppo11)
SwimmerSwimmer6 509.3 560 true final_strength=560.6 (at target!)
PointMass 863 900 true false=854, ~same

❌ Fundamental failure — Humanoid (3/25)

Env best strength target diagnosis
HumanoidRun 3.19 130 <3% target, NormalTanh distribution needed
HumanoidWalk 7.68 500 <2% target, wider policy (2×256) didn't help
HumanoidStand 18.36 700 <3% target, constant LR + wider policy tested, no improvement

Humanoid tested and failed: wider 2×256 policy + constant LR + vnorm (Wave IV-H). MA stayed flat at 8-10 for HumanoidStand over entire training. Root cause is likely NormalTanh distribution (state-dependent std + tanh squashing) — a fundamental architectural difference from Brax.


Spec Fixes Required

Priority 1: Humanoid loco spec (update ppo_playground_loco)

Official uses num_envs=8192, time_horizon=20 (unroll_length) for loco. We use num_envs=2048, time_horizon=64.

Proposed update to ppo_playground_loco:

ppo_playground_loco:
  agent:
    algorithm:
      gamma: 0.97
      time_horizon: 20      # was 64; official unroll_length=20
      training_epoch: 4
  env:
    num_envs: 8192          # was 2048; official loco num_envs=8192

Before launching: verify VRAM by checking if 8192 envs fits A4500 20GB. Run one Humanoid env, check dstack logs NAME --since 10m | grep -i "memory\|OOM" after 5 min.

Rerun only: HumanoidRun, HumanoidWalk, HumanoidStand (3 runs). HopperStand also uses loco spec — add if VRAM confirmed OK.

Priority 2: CartpoleSwingup regression

p5-ppo5 scored 803 ✅; p5-ppo6 scored ~641. The p5-ppo6 change was minibatch_size: 2048 (30 minibatches) vs p5-ppo5's 4096 (15 minibatches). More gradient steps per iter hurt CartpoleSwingup.

Option A: Revert ppo_playground minibatch_size from 2048→4096 (15 minibatches). Rerun only failing DM Control envs (CartpoleSwingup, CartpoleSwingupSparse, + any that need it).

Option B: Accept 641 and note the trade-off — p5-ppo6 improved other envs (CartpoleBalance 968 was already ✅).

Priority 3: FingerTurnEasy/Hard

No official override. At 570/? vs target 950, gap is large. Check:

grep -A10 "Finger" ~/.cache/uv/archive-v0/ON8dY3irQZTYI3Bok0SlC/mujoco_playground/config/dm_control_suite_params.py

May need deeper policy network [32,32,32,32] (official arch) vs our [64,64].


Tuning Principles Learned

  1. Check official per-env overrides first: dm_control_suite_params.py has discounting, action_repeat, num_updates_per_batch per env. These are canonical.

  2. action_repeat is env-level, not spec-level. Implemented in playground.py via _ACTION_REPEAT dict. PendulumSwingup→4. Add others as found.

  3. NaN loss: log_std clamp max=0.5 helps but Humanoid (21 DOF) still has many NaN skips. Rate-limited to log every 10K. If NaN dominates → spec is wrong.

  4. num_envs scales with task complexity: Cartpole/Acrobot: 2048 fine. Humanoid locomotion: needs 8192 for rollout diversity.

  5. time_horizon (unroll_length): DM Control official=30, loco official=20. Longer → more correlated rollouts → less diversity per update. Match official.

  6. Minibatch count: more minibatches = more gradient steps per batch. Can overfit or slow convergence for simpler envs. 15 minibatches (p5-ppo5) vs 30 (p5-ppo6) — the latter hurt CartpoleSwingup.

  7. Sparse reward + strength metric: strength (trajectory mean) severely penalizes sparse/delayed convergence. CartpoleBalanceSparse strength=545 but final MA=992. Resolve metric before publishing.

  8. High seed variance (consistency < 0): some seeds solve, some don't → wrong spec, not bad luck. Fix exploration (entropy_coef) or use different spec.

  9. -s overrides are silently ignored if the YAML key isn't a ${variable} placeholder. Always verify overrides took effect via logs: grep "gamma\|lr\|training_epoch" dstack logs.

  10. Loco spec failures: if loco spec gives <20 on env with target >100, the issue is almost certainly num_envs/time_horizon mismatch vs official, not a fundamental algo failure.


Code Changes This Session

Commit Change
8fe7bc76 playground.py: _ACTION_REPEAT lookup for per-env action_repeat. ppo_playground.yaml: added ppo_playground_fingerspin and ppo_playground_pendulum specs.
fb55c2f9 base.py: rate-limit NaN loss warning (every 10K skips). ppo_playground.yaml: revert log_frequency 1M→100K.
3f4ede3d BENCHMARKS.md: mark HopperHop ✅.

Resume Commands

# Setup
git pull && uv sync --no-default-groups

# Check jobs
dstack ps

# Intake a completed run
dstack logs RUN_NAME --since 6h 2>&1 | grep "trial_metrics" | tail -1
dstack logs RUN_NAME --since 6h 2>&1 | grep -iE "Uploading|benchmark-dev"

# Pull HF data
source .env && huggingface-cli download SLM-Lab/benchmark-dev \
  --local-dir hf_data/data/benchmark-dev --repo-type dataset \
  --include "data/FOLDER_NAME/*"

# Plot
uv run slm-lab plot -t "EnvName" -d hf_data/data/benchmark-dev/data -f FOLDER_NAME

# Launch PendulumSwingup (queue priority 1)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_pendulum train \
  -s env=playground/PendulumSwingup -s max_frame=100000000 -n p5-ppo6-pendulumswingup

# Launch FingerSpin canonical (queue priority 2)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_fingerspin train \
  -s env=playground/FingerSpin -s max_frame=100000000 -n p5-ppo6-fingerspin3

# Launch Humanoid loco (after updating ppo_playground_loco spec to num_envs=8192, time_horizon=20)
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground_loco train \
  -s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2

CRITICAL CORRECTION (2026-03-13) — Humanoid is DM Control, not Loco

Root cause of Humanoid failure: HumanoidRun/Walk/Stand are registered in dm_control_suite/__init__.py — they ARE DM Control envs. We incorrectly ran them with ppo_playground_loco (gamma=0.97, 4 epochs, time_horizon=64).

Official config uses DEFAULT DM Control params for them: discounting=0.995, 2048 envs, lr=1e-3, unroll_length=30, 16 epochs.

NaN was never the root cause — intake-b confirmed NaN skips were 0, 0, 2 in the loco runs. The spec was simply wrong.

Fix: Run all 3 Humanoid envs with ppo_playground (DM Control spec). No spec change needed.

# Launch with correct spec
source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidRun -s max_frame=100000000 -n p5-ppo6-humanoidrun2

source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidWalk -s max_frame=100000000 -n p5-ppo6-humanoidwalk2

source .env && uv run slm-lab run-remote --gpu \
  slm_lab/spec/benchmark_arc/ppo/ppo_playground.yaml ppo_playground train \
  -s env=playground/HumanoidStand -s max_frame=100000000 -n p5-ppo6-humanoidstand2

HopperStand: Also a DM Control env. If p5-ppo6-hopperstand (loco spec, 16.38) is below target, rerun with ppo_playground.

Do NOT intake the loco-spec Humanoid runs (2.78/6.82/12.45) — wrong spec, not valid benchmark results. The old ppo_playground runs (2.86/3.73) were also wrong spec but at least the right family.

Updated queue (prepend these as highest priority):

Priority Env Spec Run name
0 HumanoidRun ppo_playground p5-ppo6-humanoidrun2
0 HumanoidWalk ppo_playground p5-ppo6-humanoidwalk2
0 HumanoidStand ppo_playground p5-ppo6-humanoidstand2
0 HopperStand ppo_playground p5-ppo6-hopperstand2 (if loco result ⚠️)

Note on loco spec (ppo_playground_loco): only for actual locomotion robot envs (Go1, G1, BerkeleyHumanoid, etc.) — NOT for DM Control Humanoid.


METRIC CORRECTION (2026-03-13) — strength vs final_strength

Problem: strength = trajectory-averaged mean over entire run. For slow-rising envs this severely underrepresents end-of-training performance. After metric correction to strength:

Env strength total_reward_ma target conclusion
CartpoleSwingup 443.0 641.51 800 Massive regression from p5-ppo5 (803). Strength 443 << 665 (65M result) — curve rises but slow start drags average down
CartpoleBalanceSparse 545.1 991.81 700 Hits target by end (final MA=992) but sparse reward delays convergence
AcrobotSwingup 172.8 253.24 220 Below target by strength, above by final MA
CartpoleSwingupSparse 270.9 331.23 425 Below both metrics

Resolution needed: Reference scores from mujoco_playground are end-of-training values, not trajectory averages. final_strength (= last eval MA) is the correct comparison metric. Recommend switching BENCHMARKS.md score column to final_strength and audit all existing entries.

CartpoleSwingup regression is real regardless of metric: p5-ppo5 final_strength would be ~800+, p5-ppo6 total_reward_ma=641. The p5-ppo6 minibatch change (2048→30 minibatches) hurt CartpoleSwingup convergence speed. Fix: revert ppo_playground minibatch_size to 4096 (15 minibatches) — OR accept and investigate if CartpoleSwingup needs its own spec variant.


Next Architectural Changes

Research-based prioritized list of changes NOT yet tested. Ordered by expected impact across the most envs. Wave I (5-layer value + no grad clip) is currently running — results pending.

Priority 1: NormalTanhDistribution (tanh-squashed actions)

Expected impact: HIGH — affects FingerTurnEasy/Hard, FishSwim, Humanoid, CartpoleSwingup Implementation complexity: MEDIUM (new distribution class + policy_util changes) Envs helped: All continuous-action envs, especially precision/manipulation tasks

What Brax does differently: Brax uses NormalTanhDistribution — samples from Normal(loc, scale), then applies tanh to bound actions to [-1, 1]. The log-probability includes a log-det-jacobian correction: log_prob -= log(1 - tanh(x)^2). The scale is parameterized as softplus(raw_scale) + 0.001 (state-dependent, output by the network).

What SLM-Lab does: Raw Normal(loc, scale) with state-independent log_std as an nn.Parameter. Actions can exceed [-1, 1] and are silently clipped by the environment. The log-prob does NOT account for this clipping, creating a mismatch between the distribution the agent thinks it's using and the effective action distribution.

Why this matters:

  1. Gradient quality: Without jacobian correction, the policy gradient is biased. Actions near the boundary (common in precise manipulation like FingerTurn) have incorrect log-prob gradients. The agent cannot learn fine boundary control.
  2. Exploration: State-dependent std allows the agent to be precise where it's confident and exploratory where uncertain. State-independent std forces uniform exploration across all states — wasteful for tasks requiring both coarse and fine control.
  3. FingerTurn gap (571/950 = 60%): FingerTurn requires precise angular positioning of a fingertip. Without tanh squashing, actions at the boundary are clipped but the log-prob doesn't reflect this — the policy "thinks" it's outputting different actions that are actually identical after clipping. This prevents learning fine-grained control near action limits.
  4. Humanoid gap (<3%): 21 DOF with high-dimensional action space. State-independent std means all joints explore equally. Humanoid needs to stabilize torso (low variance) while exploring leg movement (high variance) — impossible with shared std.

Implementation plan:

  1. Add NormalTanhDistribution class in slm_lab/lib/distribution.py:
    • Forward: action = tanh(Normal(loc, scale).rsample())
    • log_prob: Normal.log_prob(atanh(action)) - log(1 - action^2 + eps)
    • entropy: approximate (no closed form for tanh-Normal)
  2. Modify policy_util.init_action_pd() to handle the new distribution
  3. Remove log_std_init for playground specs — let the network output both mean and std (state-dependent)
  4. Network change: policy output dim doubles (mean + raw_scale per action dim)

Risk: Medium. Tanh squashing changes gradient dynamics significantly. Need to validate on already-solved envs (CartpoleBalance, WalkerRun) to ensure no regression. Can gate behind a spec flag (action_pdtype: NormalTanh).


Fix 6: Constant LR variants + Humanoid variant (commit pending)

Added three new spec variants to ppo_playground.yaml:

  • ppo_playground_constlr: DM Control + constant LR + minibatch_size=4096. For envs where vnorm=false works.
  • ppo_playground_vnorm_constlr: DM Control + vnorm + constant LR + minibatch_size=2048. For precision envs.
  • ppo_playground_humanoid: 2×256 policy + constant LR + vnorm. For Humanoid DM Control envs.

Priority 2: Constant LR (remove LinearToMin decay)

Expected impact: MEDIUM — affects all envs, especially long-training ones Implementation complexity: TRIVIAL (spec-only change) Envs helped: CartpoleSwingup, CartpoleSwingupSparse, FingerTurnEasy/Hard, FishSwim

What Brax does: Constant LR = 1e-3 for all DM Control envs. No decay.

What SLM-Lab does: LinearToMin decay from 1e-3 to 3.3e-5 (min_factor=0.033) over the full training run.

Why this matters: By the midpoint of training, SLM-Lab's LR is already at ~5e-4 — half the Brax LR. By 75% of training, it's at ~2.7e-4. For envs that converge late (CartpoleSwingup, FishSwim), the LR is too low during the critical learning phase. Brax maintains full learning capacity throughout.

This was tested as part of the Brax hyperparameter bundle (Wave E) which was ALL worse, but that test changed 4 things simultaneously (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10). The constant LR was never tested in isolation.

Implementation: Set min_factor: 1.0 in spec (or remove lr_scheduler_spec entirely).

Risk: Low. Constant LR is the Brax default and widely used. If instability occurs late in training, a gentler decay (min_factor: 0.3) can be used as fallback.


Priority 3: Clip epsilon 0.3 (from 0.2)

Expected impact: MEDIUM — affects all envs Implementation complexity: TRIVIAL (spec-only change) Envs helped: FingerTurnEasy/Hard, FishSwim, CartpoleSwingup (tasks needing faster policy adaptation)

What Brax does: clipping_epsilon=0.3 for DM Control.

What SLM-Lab does: clip_eps=0.2.

Why this matters: Clip epsilon 0.2 constrains the policy ratio to [0.8, 1.2]. At 0.3, it's [0.7, 1.3] — allowing 50% larger policy updates per step. For envs that need to explore widely before converging (FingerTurn, FishSwim), the tighter constraint slows learning.

This was tested in the Brax bundle (Wave E) alongside 3 other changes — all worse together. Never tested in isolation or with just constant LR.

Implementation: Change start_val: 0.2 to start_val: 0.3 in clip_eps_spec.

Risk: Low-medium. Larger clip_eps can cause training instability with small batches. However, with our 61K batch (2048 envs * 30 steps), it should be safe. If combined with constant LR (#2), the compounding effect should be tested carefully.


Priority 4: Per-env tuning for FingerTurn (if P1-P3 insufficient)

Expected impact: HIGH for FingerTurn specifically Implementation complexity: LOW (spec variant) Envs helped: FingerTurnEasy, FingerTurnHard only

If NormalTanh + constant LR + clip_eps=0.3 don't close the FingerTurn gap (currently 60% and 51% of target), try:

  1. Lower gamma (0.99 → 0.95): FingerSpin uses gamma=0.95 officially. FingerTurn may benefit from shorter horizon discounting since reward is instantaneous (current angle vs target). Lower gamma reduces value function complexity.

  2. Smaller policy network: Brax DM Control uses (32, 32, 32, 32) — our (64, 64) may over-parameterize for manipulation tasks. Try (32, 32, 32, 32) to match exactly.

  3. Higher entropy coefficient: FingerTurn has a narrow solution manifold. Increasing entropy from 0.01 to 0.02 would encourage broader exploration of finger positions.


Priority 5: Humanoid-specific — num_envs=8192

Expected impact: HIGH for Humanoid specifically Implementation complexity: TRIVIAL (spec-only) Envs helped: HumanoidStand, HumanoidWalk, HumanoidRun

Current situation: Humanoid was incorrectly run with loco spec (gamma=0.97, 4 epochs). The correction to DM Control spec (gamma=0.995, 16 epochs) is being tested in Wave I (p5-ppo13). However, even with correct spec, the standard 2048 envs may be insufficient.

Why num_envs matters for Humanoid: 21 DOF, 67-dim observations. With 2048 envs and time_horizon=30, the batch is 61K transitions — each containing a narrow slice of the 21-DOF state space. Humanoid needs more diverse rollouts to learn coordinated multi-joint control. Brax's effective batch of 983K transitions provides 16x more state-space coverage per update.

Since we can't easily get 16x more data per update, increasing num_envs from 2048 to 4096 or 8192 doubles/quadruples rollout diversity. Combined with NormalTanh (state-dependent std for per-joint exploration), this could be sufficient.

VRAM concern: 8192 envs may exceed A4500 20GB. Test with a quick 1M frame run first. Fallback: 4096 envs.


NOT recommended (already tested, no benefit)

Change Wave Result Why it failed
normalize_v_targets: false G/G2 Mixed (helps 5, hurts 6) Already per-env split in spec
Multi-unroll (num_unrolls=16) F Same or worse by strength Stale old_net (480 vs 30 steps between copies)
Brax hyperparameter bundle (clip_eps=0.3 + constant LR + 5-layer value + reward_scale=10) E All worse Confounded — 4 changes at once. Individual effects unknown except for reward_scale (helps)
time_horizon=480 (single long unroll) C Helps CartpoleSwingup, hurts FingerTurn 480-step GAE is noisy for precision tasks
5-layer value + no grad clip III Only helped CartpoleSwingup slightly Hurt AcrobotSwingup, FishSwim; not general
NormalTanh distribution II Abandoned Architecturally incompatible — SLM-Lab stores post-tanh actions, atanh inversion unstable
vnorm=true rerun (reverted spec) IV All worse or same No new information — variance rerun
4×32 Brax policy + constant LR + vnorm VI All worse FingerTurnEasy 408 (vs 571), FingerTurnHard 244 (vs 484), FishSwim 106 (vs 463)
Humanoid wider 2×256 + constant LR + vnorm IV-H No improvement MA flat at 8-10 for all 3 Humanoid envs; NormalTanh is root cause

Currently testing

Wave V-B completed results (constant LR)

Env strength final_strength Old best Verdict
PointMass 841.3 877.3 863.5 ❌ strength lower
SwimmerSwimmer6 517.3 585.7 509.3 ✅ NEW BEST (+1.6%)
FishSwim 434.6 550.8 463.0 ❌ strength lower (final much better)

Wave VII completed results (clip_eps=0.3 + constant LR)

Env strength final_strength Old best Verdict
FingerTurnEasy 518.0 608.8 570.9 ❌ strength lower (final much better, but slow start drags average)
FingerTurnHard 401.7 489.7 484.1 ❌ strength lower (same pattern)
FishSwim 476.9 581.4 463.0 ✅ NEW BEST (+3%)

Key insight: clip_eps=0.3 produces higher final performance but worse trajectory-averaged strength. The wider clip allows bigger policy updates which increases exploration early (slower convergence) but reaches higher asymptotic performance. The strength metric penalizes late bloomers.

Wave V completed results

Env strength final_strength Old best Verdict
CartpoleSwingup 606.5 702.6 576.1 ✅ NEW BEST (+5%)
CartpoleSwingupSparse 383.7 536.2 296.3 ✅ NEW BEST (+29%)
CartpoleBalanceSparse 757.9 993.0 690.4 ✅ NEW BEST (+10%)
AcrobotSwingup 161.2 246.9 172.8 ❌ strength lower (final_strength much better but trajectory avg worse due to slow start)

Key insight: Constant LR is the single most impactful change found. LR decay from 1e-3 to 3.3e-5 was hurting late-converging envs. CartpoleBalanceSparse went from 690→993 (final_strength), effectively solved.

Completed waves

Wave VI (p5-ppo18): 4×32 Brax policy — STOPPED, all underperformed. FingerTurnEasy MA 408, FingerTurnHard MA 244, FishSwim MA 106. All below old bests.

Wave IV-H (p5-ppo16h): Humanoid wider 2×256 + constant LR + vnorm — all flat at MA 8-10.

Next steps after Wave VII

  1. Humanoid num_envs=4096/8192 — only major gap remaining after Wave VII
  2. Consider constant LR + clip_eps=0.3 as new general default if results hold across envs

Key Brax architecture differences (from source code analysis)

Parameter Brax Default SLM-Lab Impact
Policy 4×32 (deeper, narrower) 2×64 Testable via spec
Value 5×256 3×256 Tested Wave III — no help
Distribution tanh_normal Normal Cannot test (architectural incompatibility)
Init lecun_uniform orthogonal_ Would need code change
State-dep std False (scalar) False (nn.Parameter) Similar
Activation swish (SiLU) SiLU ✅ Match
clipping_epsilon 0.3 0.2 Testable via spec
num_minibatches 32 15-30 Close enough
num_unrolls 16 (implicit) 1 Tested Wave F — stale old_net hurts