pick-place-v2-sf / sf_log.txt
qgallouedec's picture
qgallouedec HF staff
Upload . with huggingface_hub
63244f8
raw
history blame
91.6 kB
[2023-03-03 11:18:30,885][16922] Saving configuration to /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/config.json...
[2023-03-03 11:18:30,909][16922] Rollout worker 0 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 1 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 2 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 3 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 4 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 5 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 6 uses device cpu
[2023-03-03 11:18:30,910][16922] Rollout worker 7 uses device cpu
[2023-03-03 11:19:39,471][16994] Saving configuration to /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/config.json...
[2023-03-03 11:19:39,496][16994] Rollout worker 0 uses device cpu
[2023-03-03 11:19:39,496][16994] Rollout worker 1 uses device cpu
[2023-03-03 11:19:39,496][16994] Rollout worker 2 uses device cpu
[2023-03-03 11:19:39,497][16994] Rollout worker 3 uses device cpu
[2023-03-03 11:19:39,497][16994] Rollout worker 4 uses device cpu
[2023-03-03 11:19:39,497][16994] Rollout worker 5 uses device cpu
[2023-03-03 11:19:39,497][16994] Rollout worker 6 uses device cpu
[2023-03-03 11:19:39,497][16994] Rollout worker 7 uses device cpu
[2023-03-03 11:19:39,675][16994] InferenceWorker_p0-w0: min num requests: 2
[2023-03-03 11:19:39,713][16994] Starting all processes...
[2023-03-03 11:19:39,714][16994] Starting process learner_proc0
[2023-03-03 11:19:39,772][16994] Starting all processes...
[2023-03-03 11:19:39,793][16994] Starting process inference_proc0-0
[2023-03-03 11:19:39,799][16994] Starting process rollout_proc0
[2023-03-03 11:19:39,800][16994] Starting process rollout_proc1
[2023-03-03 11:19:39,804][16994] Starting process rollout_proc2
[2023-03-03 11:19:39,813][16994] Starting process rollout_proc3
[2023-03-03 11:19:39,817][16994] Starting process rollout_proc4
[2023-03-03 11:19:39,826][16994] Starting process rollout_proc5
[2023-03-03 11:19:39,827][16994] Starting process rollout_proc6
[2023-03-03 11:19:39,828][16994] Starting process rollout_proc7
[2023-03-03 11:19:43,526][17030] WARNING! It is generally recommended to enable Fixed KL loss (https://arxiv.org/pdf/1707.06347.pdf) for continuous action tasks to avoid potential numerical issues. I.e. set --kl_loss_coeff=0.1
[2023-03-03 11:19:43,526][17030] Starting seed is not provided
[2023-03-03 11:19:43,526][17030] Initializing actor-critic model on device cpu
[2023-03-03 11:19:43,526][17030] RunningMeanStd input shape: (39,)
[2023-03-03 11:19:43,528][17030] RunningMeanStd input shape: (1,)
[2023-03-03 11:19:43,570][17037] On MacOS, not setting affinity
[2023-03-03 11:19:43,571][17039] On MacOS, not setting affinity
[2023-03-03 11:19:43,666][17034] On MacOS, not setting affinity
[2023-03-03 11:19:43,669][17030] Created Actor Critic model with architecture:
[2023-03-03 11:19:43,669][17030] ActorCriticSharedWeights(
(obs_normalizer): ObservationNormalizer(
(running_mean_std): RunningMeanStdDictInPlace(
(running_mean_std): ModuleDict(
(obs): RunningMeanStdInPlace()
)
)
)
(returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
(encoder): MultiInputEncoder(
(encoders): ModuleDict(
(obs): MlpEncoder(
(mlp_head): RecursiveScriptModule(
original_name=Sequential
(0): RecursiveScriptModule(original_name=Linear)
(1): RecursiveScriptModule(original_name=ELU)
(2): RecursiveScriptModule(original_name=Linear)
(3): RecursiveScriptModule(original_name=ELU)
)
)
)
)
(core): ModelCoreRNN(
(core): GRU(512, 512)
)
(decoder): MlpDecoder(
(mlp): Identity()
)
(critic_linear): Linear(in_features=512, out_features=1, bias=True)
(action_parameterization): ActionParameterizationDefault(
(distribution_linear): Linear(in_features=512, out_features=8, bias=True)
)
)
[2023-03-03 11:19:43,686][17030] Using optimizer <class 'torch.optim.adam.Adam'>
[2023-03-03 11:19:43,687][17030] No checkpoints found
[2023-03-03 11:19:43,688][17030] Did not load from checkpoint, starting from scratch!
[2023-03-03 11:19:43,693][17030] Initialized policy 0 weights for model version 0
[2023-03-03 11:19:43,694][17030] LearnerWorker_p0 finished initialization!
[2023-03-03 11:19:43,752][17038] On MacOS, not setting affinity
[2023-03-03 11:19:43,813][17032] On MacOS, not setting affinity
[2023-03-03 11:19:43,816][17031] RunningMeanStd input shape: (39,)
[2023-03-03 11:19:43,817][17031] RunningMeanStd input shape: (1,)
[2023-03-03 11:19:43,835][17036] On MacOS, not setting affinity
[2023-03-03 11:19:43,846][17033] On MacOS, not setting affinity
[2023-03-03 11:19:43,846][17035] On MacOS, not setting affinity
[2023-03-03 11:19:43,875][16994] Inference worker 0-0 is ready!
[2023-03-03 11:19:43,877][16994] All inference workers are ready! Signal rollout workers to start!
[2023-03-03 11:19:44,848][17039] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,883][17037] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,883][17032] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,932][17038] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,935][17034] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,966][17035] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,969][17033] Decorrelating experience for 0 frames...
[2023-03-03 11:19:44,995][17036] Decorrelating experience for 0 frames...
[2023-03-03 11:19:45,824][17037] Decorrelating experience for 32 frames...
[2023-03-03 11:19:45,846][17032] Decorrelating experience for 32 frames...
[2023-03-03 11:19:45,898][17038] Decorrelating experience for 32 frames...
[2023-03-03 11:19:45,938][17034] Decorrelating experience for 32 frames...
[2023-03-03 11:19:45,946][17033] Decorrelating experience for 32 frames...
[2023-03-03 11:19:45,992][17036] Decorrelating experience for 32 frames...
[2023-03-03 11:19:46,027][17035] Decorrelating experience for 32 frames...
[2023-03-03 11:19:46,047][17039] Decorrelating experience for 32 frames...
[2023-03-03 11:19:47,683][16994] Fps is (10 sec: nan, 60 sec: nan, 300 sec: nan). Total num frames: 4096. Throughput: 0: nan. Samples: 2914. Policy #0 lag: (min: 0.0, avg: 0.0, max: 0.0)
[2023-03-03 11:19:49,484][17031] Updated weights for policy 0, policy_version 10 (0.0008)
[2023-03-03 11:19:52,621][17031] Updated weights for policy 0, policy_version 20 (0.0007)
[2023-03-03 11:19:52,682][16994] Fps is (10 sec: 3277.0, 60 sec: 3277.0, 300 sec: 3277.0). Total num frames: 20480. Throughput: 0: 1888.9. Samples: 12358. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:19:52,683][16994] Avg episode reward: [(0, '4.341')]
[2023-03-03 11:19:55,978][17031] Updated weights for policy 0, policy_version 30 (0.0008)
[2023-03-03 11:19:57,684][16994] Fps is (10 sec: 3174.0, 60 sec: 3174.0, 300 sec: 3174.0). Total num frames: 35840. Throughput: 0: 2828.4. Samples: 31202. Policy #0 lag: (min: 0.0, avg: 0.6, max: 1.0)
[2023-03-03 11:19:57,685][16994] Avg episode reward: [(0, '5.303')]
[2023-03-03 11:19:59,217][17031] Updated weights for policy 0, policy_version 40 (0.0006)
[2023-03-03 11:19:59,664][16994] Heartbeat connected on Batcher_0
[2023-03-03 11:19:59,681][16994] Heartbeat connected on InferenceWorker_p0-w0
[2023-03-03 11:19:59,683][16994] Heartbeat connected on RolloutWorker_w0
[2023-03-03 11:19:59,687][16994] Heartbeat connected on RolloutWorker_w1
[2023-03-03 11:19:59,691][16994] Heartbeat connected on RolloutWorker_w2
[2023-03-03 11:19:59,696][16994] Heartbeat connected on RolloutWorker_w3
[2023-03-03 11:19:59,702][16994] Heartbeat connected on RolloutWorker_w4
[2023-03-03 11:19:59,707][16994] Heartbeat connected on RolloutWorker_w5
[2023-03-03 11:19:59,718][16994] Heartbeat connected on RolloutWorker_w6
[2023-03-03 11:19:59,721][16994] Heartbeat connected on RolloutWorker_w7
[2023-03-03 11:19:59,864][16994] Heartbeat connected on LearnerWorker_p0
[2023-03-03 11:20:02,508][17031] Updated weights for policy 0, policy_version 50 (0.0008)
[2023-03-03 11:20:02,681][16994] Fps is (10 sec: 3072.4, 60 sec: 3140.6, 300 sec: 3140.6). Total num frames: 51200. Throughput: 0: 3133.0. Samples: 49904. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:20:02,682][16994] Avg episode reward: [(0, '5.930')]
[2023-03-03 11:20:05,680][17031] Updated weights for policy 0, policy_version 60 (0.0007)
[2023-03-03 11:20:07,681][16994] Fps is (10 sec: 3175.4, 60 sec: 3174.7, 300 sec: 3174.7). Total num frames: 67584. Throughput: 0: 2827.8. Samples: 59466. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:20:07,681][16994] Avg episode reward: [(0, '6.461')]
[2023-03-03 11:20:08,979][17031] Updated weights for policy 0, policy_version 70 (0.0006)
[2023-03-03 11:20:12,410][17031] Updated weights for policy 0, policy_version 80 (0.0009)
[2023-03-03 11:20:12,685][16994] Fps is (10 sec: 3070.9, 60 sec: 3112.7, 300 sec: 3112.7). Total num frames: 81920. Throughput: 0: 3011.7. Samples: 78213. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:20:12,688][16994] Avg episode reward: [(0, '6.726')]
[2023-03-03 11:20:16,059][17031] Updated weights for policy 0, policy_version 90 (0.0009)
[2023-03-03 11:20:17,682][16994] Fps is (10 sec: 2764.5, 60 sec: 3037.9, 300 sec: 3037.9). Total num frames: 95232. Throughput: 0: 3050.6. Samples: 94430. Policy #0 lag: (min: 0.0, avg: 0.6, max: 1.0)
[2023-03-03 11:20:17,685][16994] Avg episode reward: [(0, '6.601')]
[2023-03-03 11:20:19,878][17031] Updated weights for policy 0, policy_version 100 (0.0007)
[2023-03-03 11:20:22,682][16994] Fps is (10 sec: 2765.4, 60 sec: 3013.5, 300 sec: 3013.5). Total num frames: 109568. Throughput: 0: 2863.2. Samples: 103126. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:20:22,685][16994] Avg episode reward: [(0, '6.769')]
[2023-03-03 11:20:22,734][17030] Saving new best policy, reward=6.769!
[2023-03-03 11:20:23,564][17031] Updated weights for policy 0, policy_version 110 (0.0008)
[2023-03-03 11:20:26,793][17031] Updated weights for policy 0, policy_version 120 (0.0008)
[2023-03-03 11:20:27,684][16994] Fps is (10 sec: 2969.1, 60 sec: 3020.7, 300 sec: 3020.7). Total num frames: 124928. Throughput: 0: 2944.1. Samples: 120681. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:20:27,685][16994] Avg episode reward: [(0, '7.823')]
[2023-03-03 11:20:27,751][17030] Saving new best policy, reward=7.823!
[2023-03-03 11:20:30,017][17031] Updated weights for policy 0, policy_version 130 (0.0006)
[2023-03-03 11:20:32,683][16994] Fps is (10 sec: 3174.2, 60 sec: 3049.2, 300 sec: 3049.2). Total num frames: 141312. Throughput: 0: 3038.5. Samples: 139649. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:20:32,684][16994] Avg episode reward: [(0, '8.973')]
[2023-03-03 11:20:32,684][17030] Saving new best policy, reward=8.973!
[2023-03-03 11:20:33,250][17031] Updated weights for policy 0, policy_version 140 (0.0007)
[2023-03-03 11:20:36,407][17031] Updated weights for policy 0, policy_version 150 (0.0007)
[2023-03-03 11:20:37,682][16994] Fps is (10 sec: 3277.3, 60 sec: 3072.0, 300 sec: 3072.0). Total num frames: 157696. Throughput: 0: 3039.4. Samples: 149131. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:20:37,683][16994] Avg episode reward: [(0, '10.288')]
[2023-03-03 11:20:37,688][17030] Saving new best policy, reward=10.288!
[2023-03-03 11:20:39,529][17031] Updated weights for policy 0, policy_version 160 (0.0007)
[2023-03-03 11:20:42,684][16994] Fps is (10 sec: 3071.8, 60 sec: 3053.3, 300 sec: 3053.3). Total num frames: 172032. Throughput: 0: 3028.3. Samples: 167475. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:20:42,685][16994] Avg episode reward: [(0, '10.363')]
[2023-03-03 11:20:42,686][17030] Saving new best policy, reward=10.363!
[2023-03-03 11:20:43,362][17031] Updated weights for policy 0, policy_version 170 (0.0009)
[2023-03-03 11:20:46,966][17031] Updated weights for policy 0, policy_version 180 (0.0007)
[2023-03-03 11:20:47,683][16994] Fps is (10 sec: 2867.2, 60 sec: 3037.9, 300 sec: 3037.9). Total num frames: 186368. Throughput: 0: 3000.2. Samples: 184916. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:20:47,685][16994] Avg episode reward: [(0, '11.854')]
[2023-03-03 11:20:47,689][17030] Saving new best policy, reward=11.854!
[2023-03-03 11:20:51,032][17031] Updated weights for policy 0, policy_version 190 (0.0008)
[2023-03-03 11:20:52,683][16994] Fps is (10 sec: 2560.2, 60 sec: 2952.5, 300 sec: 2977.5). Total num frames: 197632. Throughput: 0: 2959.4. Samples: 192647. Policy #0 lag: (min: 0.0, avg: 0.6, max: 2.0)
[2023-03-03 11:20:52,686][16994] Avg episode reward: [(0, '11.889')]
[2023-03-03 11:20:52,719][17030] Saving new best policy, reward=11.889!
[2023-03-03 11:20:54,924][17031] Updated weights for policy 0, policy_version 200 (0.0008)
[2023-03-03 11:20:57,683][16994] Fps is (10 sec: 2662.3, 60 sec: 2952.6, 300 sec: 2984.2). Total num frames: 212992. Throughput: 0: 2898.2. Samples: 208629. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:20:57,685][16994] Avg episode reward: [(0, '13.343')]
[2023-03-03 11:20:57,691][17030] Saving new best policy, reward=13.343!
[2023-03-03 11:20:58,304][17031] Updated weights for policy 0, policy_version 210 (0.0013)
[2023-03-03 11:21:02,239][17031] Updated weights for policy 0, policy_version 220 (0.0010)
[2023-03-03 11:21:02,683][16994] Fps is (10 sec: 2867.1, 60 sec: 2918.3, 300 sec: 2962.8). Total num frames: 226304. Throughput: 0: 2909.3. Samples: 225350. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:21:02,685][16994] Avg episode reward: [(0, '14.866')]
[2023-03-03 11:21:02,686][17030] Saving new best policy, reward=14.866!
[2023-03-03 11:21:05,679][17031] Updated weights for policy 0, policy_version 230 (0.0007)
[2023-03-03 11:21:07,683][16994] Fps is (10 sec: 2764.7, 60 sec: 2884.1, 300 sec: 2956.8). Total num frames: 240640. Throughput: 0: 2904.5. Samples: 233833. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:21:07,684][16994] Avg episode reward: [(0, '15.002')]
[2023-03-03 11:21:07,767][17030] Saving new best policy, reward=15.002!
[2023-03-03 11:21:09,100][17031] Updated weights for policy 0, policy_version 240 (0.0007)
[2023-03-03 11:21:12,427][17031] Updated weights for policy 0, policy_version 250 (0.0009)
[2023-03-03 11:21:12,682][16994] Fps is (10 sec: 2970.0, 60 sec: 2901.5, 300 sec: 2963.6). Total num frames: 256000. Throughput: 0: 2922.0. Samples: 252164. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:21:12,683][16994] Avg episode reward: [(0, '15.373')]
[2023-03-03 11:21:12,748][17030] Saving new best policy, reward=15.373!
[2023-03-03 11:21:15,846][17031] Updated weights for policy 0, policy_version 260 (0.0007)
[2023-03-03 11:21:17,685][16994] Fps is (10 sec: 2969.2, 60 sec: 2918.3, 300 sec: 2958.2). Total num frames: 270336. Throughput: 0: 2890.5. Samples: 269724. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:21:17,687][16994] Avg episode reward: [(0, '14.800')]
[2023-03-03 11:21:19,480][17031] Updated weights for policy 0, policy_version 270 (0.0009)
[2023-03-03 11:21:22,684][16994] Fps is (10 sec: 2866.6, 60 sec: 2918.3, 300 sec: 2953.4). Total num frames: 284672. Throughput: 0: 2871.5. Samples: 278353. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:21:22,687][16994] Avg episode reward: [(0, '14.436')]
[2023-03-03 11:21:23,184][17031] Updated weights for policy 0, policy_version 280 (0.0007)
[2023-03-03 11:21:27,106][17031] Updated weights for policy 0, policy_version 290 (0.0008)
[2023-03-03 11:21:27,684][16994] Fps is (10 sec: 2764.9, 60 sec: 2884.2, 300 sec: 2938.8). Total num frames: 297984. Throughput: 0: 2818.6. Samples: 294315. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:21:27,687][16994] Avg episode reward: [(0, '14.826')]
[2023-03-03 11:21:30,845][17031] Updated weights for policy 0, policy_version 300 (0.0007)
[2023-03-03 11:21:32,685][16994] Fps is (10 sec: 2559.8, 60 sec: 2815.9, 300 sec: 2915.9). Total num frames: 310272. Throughput: 0: 2779.1. Samples: 309983. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:21:32,688][16994] Avg episode reward: [(0, '15.389')]
[2023-03-03 11:21:32,695][17030] Saving new best policy, reward=15.389!
[2023-03-03 11:21:34,888][17031] Updated weights for policy 0, policy_version 310 (0.0010)
[2023-03-03 11:21:37,685][16994] Fps is (10 sec: 2662.3, 60 sec: 2781.8, 300 sec: 2913.7). Total num frames: 324608. Throughput: 0: 2787.5. Samples: 318090. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:21:37,686][16994] Avg episode reward: [(0, '16.356')]
[2023-03-03 11:21:37,835][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000000318_325632.pth...
[2023-03-03 11:21:37,925][17030] Saving new best policy, reward=16.356!
[2023-03-03 11:21:38,517][17031] Updated weights for policy 0, policy_version 320 (0.0008)
[2023-03-03 11:21:42,684][16994] Fps is (10 sec: 2662.6, 60 sec: 2747.7, 300 sec: 2893.9). Total num frames: 336896. Throughput: 0: 2790.7. Samples: 334213. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:21:42,686][16994] Avg episode reward: [(0, '18.428')]
[2023-03-03 11:21:42,687][17030] Saving new best policy, reward=18.428!
[2023-03-03 11:21:42,949][17031] Updated weights for policy 0, policy_version 330 (0.0018)
[2023-03-03 11:21:46,353][17031] Updated weights for policy 0, policy_version 340 (0.0007)
[2023-03-03 11:21:47,683][16994] Fps is (10 sec: 2662.9, 60 sec: 2747.7, 300 sec: 2892.8). Total num frames: 351232. Throughput: 0: 2781.5. Samples: 350515. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:21:47,686][16994] Avg episode reward: [(0, '20.106')]
[2023-03-03 11:21:47,831][17030] Saving new best policy, reward=20.106!
[2023-03-03 11:21:50,300][17031] Updated weights for policy 0, policy_version 350 (0.0008)
[2023-03-03 11:21:52,684][16994] Fps is (10 sec: 2867.1, 60 sec: 2798.9, 300 sec: 2891.7). Total num frames: 365568. Throughput: 0: 2760.1. Samples: 358039. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:21:52,687][16994] Avg episode reward: [(0, '20.657')]
[2023-03-03 11:21:52,795][17030] Saving new best policy, reward=20.657!
[2023-03-03 11:21:53,533][17031] Updated weights for policy 0, policy_version 360 (0.0008)
[2023-03-03 11:21:56,808][17031] Updated weights for policy 0, policy_version 370 (0.0007)
[2023-03-03 11:21:57,681][16994] Fps is (10 sec: 2970.2, 60 sec: 2799.0, 300 sec: 2898.7). Total num frames: 380928. Throughput: 0: 2761.0. Samples: 376406. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:21:57,681][16994] Avg episode reward: [(0, '21.310')]
[2023-03-03 11:21:57,689][17030] Saving new best policy, reward=21.310!
[2023-03-03 11:22:00,298][17031] Updated weights for policy 0, policy_version 380 (0.0008)
[2023-03-03 11:22:02,683][16994] Fps is (10 sec: 3072.3, 60 sec: 2833.1, 300 sec: 2905.1). Total num frames: 396288. Throughput: 0: 2778.3. Samples: 394744. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:22:02,684][16994] Avg episode reward: [(0, '22.168')]
[2023-03-03 11:22:02,685][17030] Saving new best policy, reward=22.168!
[2023-03-03 11:22:03,531][17031] Updated weights for policy 0, policy_version 390 (0.0008)
[2023-03-03 11:22:06,890][17031] Updated weights for policy 0, policy_version 400 (0.0008)
[2023-03-03 11:22:07,683][16994] Fps is (10 sec: 3071.2, 60 sec: 2850.1, 300 sec: 2911.1). Total num frames: 411648. Throughput: 0: 2796.3. Samples: 404187. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:22:07,684][16994] Avg episode reward: [(0, '21.111')]
[2023-03-03 11:22:10,690][17031] Updated weights for policy 0, policy_version 410 (0.0016)
[2023-03-03 11:22:12,683][16994] Fps is (10 sec: 2764.7, 60 sec: 2798.9, 300 sec: 2895.4). Total num frames: 423936. Throughput: 0: 2824.0. Samples: 421391. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:12,686][16994] Avg episode reward: [(0, '27.536')]
[2023-03-03 11:22:12,690][17030] Saving new best policy, reward=27.536!
[2023-03-03 11:22:14,185][17031] Updated weights for policy 0, policy_version 420 (0.0007)
[2023-03-03 11:22:17,683][16994] Fps is (10 sec: 2764.9, 60 sec: 2816.1, 300 sec: 2901.3). Total num frames: 439296. Throughput: 0: 2865.4. Samples: 438923. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:22:17,684][16994] Avg episode reward: [(0, '27.382')]
[2023-03-03 11:22:17,819][17031] Updated weights for policy 0, policy_version 430 (0.0007)
[2023-03-03 11:22:21,346][17031] Updated weights for policy 0, policy_version 440 (0.0008)
[2023-03-03 11:22:22,682][16994] Fps is (10 sec: 3072.6, 60 sec: 2833.2, 300 sec: 2906.9). Total num frames: 454656. Throughput: 0: 2881.6. Samples: 447751. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:22,697][16994] Avg episode reward: [(0, '28.442')]
[2023-03-03 11:22:22,867][17030] Saving new best policy, reward=28.442!
[2023-03-03 11:22:25,095][17031] Updated weights for policy 0, policy_version 450 (0.0008)
[2023-03-03 11:22:27,684][16994] Fps is (10 sec: 2867.0, 60 sec: 2833.1, 300 sec: 2899.2). Total num frames: 467968. Throughput: 0: 2869.5. Samples: 463342. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:22:27,686][16994] Avg episode reward: [(0, '23.001')]
[2023-03-03 11:22:28,598][17031] Updated weights for policy 0, policy_version 460 (0.0007)
[2023-03-03 11:22:31,685][17031] Updated weights for policy 0, policy_version 470 (0.0007)
[2023-03-03 11:22:32,683][16994] Fps is (10 sec: 2969.0, 60 sec: 2901.4, 300 sec: 2910.6). Total num frames: 484352. Throughput: 0: 2941.1. Samples: 482865. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:32,684][16994] Avg episode reward: [(0, '23.627')]
[2023-03-03 11:22:34,951][17031] Updated weights for policy 0, policy_version 480 (0.0007)
[2023-03-03 11:22:37,684][16994] Fps is (10 sec: 3071.9, 60 sec: 2901.4, 300 sec: 2909.3). Total num frames: 498688. Throughput: 0: 2978.7. Samples: 492082. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:22:37,686][16994] Avg episode reward: [(0, '24.366')]
[2023-03-03 11:22:38,503][17031] Updated weights for policy 0, policy_version 490 (0.0008)
[2023-03-03 11:22:42,203][17031] Updated weights for policy 0, policy_version 500 (0.0010)
[2023-03-03 11:22:42,683][16994] Fps is (10 sec: 2867.3, 60 sec: 2935.5, 300 sec: 2908.2). Total num frames: 513024. Throughput: 0: 2958.1. Samples: 509528. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:22:42,685][16994] Avg episode reward: [(0, '25.101')]
[2023-03-03 11:22:45,721][17031] Updated weights for policy 0, policy_version 510 (0.0007)
[2023-03-03 11:22:47,683][16994] Fps is (10 sec: 2867.6, 60 sec: 2935.5, 300 sec: 2907.0). Total num frames: 527360. Throughput: 0: 2929.4. Samples: 526565. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:47,684][16994] Avg episode reward: [(0, '26.304')]
[2023-03-03 11:22:49,180][17031] Updated weights for policy 0, policy_version 520 (0.0009)
[2023-03-03 11:22:52,382][17031] Updated weights for policy 0, policy_version 530 (0.0008)
[2023-03-03 11:22:52,683][16994] Fps is (10 sec: 2969.5, 60 sec: 2952.6, 300 sec: 2911.5). Total num frames: 542720. Throughput: 0: 2913.9. Samples: 535311. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:52,685][16994] Avg episode reward: [(0, '27.157')]
[2023-03-03 11:22:55,888][17031] Updated weights for policy 0, policy_version 540 (0.0007)
[2023-03-03 11:22:57,681][16994] Fps is (10 sec: 3175.1, 60 sec: 2969.6, 300 sec: 2921.1). Total num frames: 559104. Throughput: 0: 2941.3. Samples: 553743. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:22:57,682][16994] Avg episode reward: [(0, '28.972')]
[2023-03-03 11:22:57,690][17030] Saving new best policy, reward=28.972!
[2023-03-03 11:22:58,986][17031] Updated weights for policy 0, policy_version 550 (0.0007)
[2023-03-03 11:23:02,056][17031] Updated weights for policy 0, policy_version 560 (0.0007)
[2023-03-03 11:23:02,683][16994] Fps is (10 sec: 3276.8, 60 sec: 2986.7, 300 sec: 2930.2). Total num frames: 575488. Throughput: 0: 2994.2. Samples: 573661. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:23:02,684][16994] Avg episode reward: [(0, '30.927')]
[2023-03-03 11:23:02,684][17030] Saving new best policy, reward=30.927!
[2023-03-03 11:23:05,153][17031] Updated weights for policy 0, policy_version 570 (0.0007)
[2023-03-03 11:23:07,682][16994] Fps is (10 sec: 3173.9, 60 sec: 2986.7, 300 sec: 2933.8). Total num frames: 590848. Throughput: 0: 3024.4. Samples: 583852. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:07,684][16994] Avg episode reward: [(0, '31.405')]
[2023-03-03 11:23:07,868][17030] Saving new best policy, reward=31.405!
[2023-03-03 11:23:08,519][17031] Updated weights for policy 0, policy_version 580 (0.0007)
[2023-03-03 11:23:12,150][17031] Updated weights for policy 0, policy_version 590 (0.0010)
[2023-03-03 11:23:12,681][16994] Fps is (10 sec: 2970.4, 60 sec: 3020.9, 300 sec: 2932.2). Total num frames: 605184. Throughput: 0: 3077.5. Samples: 601821. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:23:12,683][16994] Avg episode reward: [(0, '32.142')]
[2023-03-03 11:23:12,842][17030] Saving new best policy, reward=32.142!
[2023-03-03 11:23:15,561][17031] Updated weights for policy 0, policy_version 600 (0.0007)
[2023-03-03 11:23:17,703][16994] Fps is (10 sec: 2861.2, 60 sec: 3002.7, 300 sec: 2930.3). Total num frames: 619520. Throughput: 0: 3026.6. Samples: 619123. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:17,727][16994] Avg episode reward: [(0, '30.425')]
[2023-03-03 11:23:19,316][17031] Updated weights for policy 0, policy_version 610 (0.0009)
[2023-03-03 11:23:22,641][17031] Updated weights for policy 0, policy_version 620 (0.0010)
[2023-03-03 11:23:22,681][16994] Fps is (10 sec: 2969.5, 60 sec: 3003.8, 300 sec: 2933.9). Total num frames: 634880. Throughput: 0: 2999.8. Samples: 627064. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:22,681][16994] Avg episode reward: [(0, '31.084')]
[2023-03-03 11:23:27,305][17031] Updated weights for policy 0, policy_version 630 (0.0010)
[2023-03-03 11:23:27,690][16994] Fps is (10 sec: 2666.5, 60 sec: 2969.4, 300 sec: 2918.3). Total num frames: 646144. Throughput: 0: 2942.3. Samples: 641943. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:23:27,694][16994] Avg episode reward: [(0, '33.239')]
[2023-03-03 11:23:27,701][17030] Saving new best policy, reward=33.239!
[2023-03-03 11:23:31,042][17031] Updated weights for policy 0, policy_version 640 (0.0007)
[2023-03-03 11:23:32,682][16994] Fps is (10 sec: 2354.8, 60 sec: 2901.4, 300 sec: 2908.2). Total num frames: 658432. Throughput: 0: 2921.9. Samples: 658047. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:23:32,687][16994] Avg episode reward: [(0, '35.241')]
[2023-03-03 11:23:32,692][17030] Saving new best policy, reward=35.241!
[2023-03-03 11:23:34,624][17031] Updated weights for policy 0, policy_version 650 (0.0007)
[2023-03-03 11:23:37,683][16994] Fps is (10 sec: 2766.0, 60 sec: 2918.4, 300 sec: 2911.7). Total num frames: 673792. Throughput: 0: 2931.5. Samples: 667231. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:37,685][16994] Avg episode reward: [(0, '36.570')]
[2023-03-03 11:23:37,832][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000000659_674816.pth...
[2023-03-03 11:23:37,881][17030] Saving new best policy, reward=36.570!
[2023-03-03 11:23:38,157][17031] Updated weights for policy 0, policy_version 660 (0.0007)
[2023-03-03 11:23:41,391][17031] Updated weights for policy 0, policy_version 670 (0.0007)
[2023-03-03 11:23:42,684][16994] Fps is (10 sec: 2969.2, 60 sec: 2918.4, 300 sec: 2910.8). Total num frames: 688128. Throughput: 0: 2922.6. Samples: 685271. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:42,688][16994] Avg episode reward: [(0, '36.296')]
[2023-03-03 11:23:45,880][17031] Updated weights for policy 0, policy_version 680 (0.0010)
[2023-03-03 11:23:47,689][16994] Fps is (10 sec: 2662.2, 60 sec: 2884.2, 300 sec: 2901.3). Total num frames: 700416. Throughput: 0: 2813.8. Samples: 700286. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:23:47,699][16994] Avg episode reward: [(0, '36.345')]
[2023-03-03 11:23:49,728][17031] Updated weights for policy 0, policy_version 690 (0.0007)
[2023-03-03 11:23:52,684][16994] Fps is (10 sec: 2560.1, 60 sec: 2850.1, 300 sec: 2896.4). Total num frames: 713728. Throughput: 0: 2757.5. Samples: 707942. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:23:52,688][16994] Avg episode reward: [(0, '35.173')]
[2023-03-03 11:23:53,667][17031] Updated weights for policy 0, policy_version 700 (0.0009)
[2023-03-03 11:23:57,043][17031] Updated weights for policy 0, policy_version 710 (0.0007)
[2023-03-03 11:23:57,681][16994] Fps is (10 sec: 2765.6, 60 sec: 2816.0, 300 sec: 2895.9). Total num frames: 728064. Throughput: 0: 2725.5. Samples: 724470. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:23:57,685][16994] Avg episode reward: [(0, '36.033')]
[2023-03-03 11:24:00,257][17031] Updated weights for policy 0, policy_version 720 (0.0007)
[2023-03-03 11:24:02,683][16994] Fps is (10 sec: 3072.1, 60 sec: 2816.0, 300 sec: 2903.3). Total num frames: 744448. Throughput: 0: 2760.8. Samples: 743303. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:24:02,684][16994] Avg episode reward: [(0, '38.665')]
[2023-03-03 11:24:02,754][17030] Saving new best policy, reward=38.665!
[2023-03-03 11:24:03,571][17031] Updated weights for policy 0, policy_version 730 (0.0007)
[2023-03-03 11:24:06,817][17031] Updated weights for policy 0, policy_version 740 (0.0009)
[2023-03-03 11:24:07,684][16994] Fps is (10 sec: 3071.3, 60 sec: 2798.9, 300 sec: 2902.6). Total num frames: 758784. Throughput: 0: 2784.8. Samples: 752387. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:24:07,686][16994] Avg episode reward: [(0, '40.911')]
[2023-03-03 11:24:07,744][17030] Saving new best policy, reward=40.911!
[2023-03-03 11:24:10,557][17031] Updated weights for policy 0, policy_version 750 (0.0007)
[2023-03-03 11:24:12,682][16994] Fps is (10 sec: 2970.0, 60 sec: 2815.9, 300 sec: 2905.9). Total num frames: 774144. Throughput: 0: 2842.3. Samples: 769832. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:24:12,684][16994] Avg episode reward: [(0, '39.414')]
[2023-03-03 11:24:13,705][17031] Updated weights for policy 0, policy_version 760 (0.0007)
[2023-03-03 11:24:16,956][17031] Updated weights for policy 0, policy_version 770 (0.0006)
[2023-03-03 11:24:17,681][16994] Fps is (10 sec: 3175.2, 60 sec: 2851.2, 300 sec: 2912.7). Total num frames: 790528. Throughput: 0: 2912.3. Samples: 789099. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:24:17,682][16994] Avg episode reward: [(0, '38.017')]
[2023-03-03 11:24:20,225][17031] Updated weights for policy 0, policy_version 780 (0.0007)
[2023-03-03 11:24:22,683][16994] Fps is (10 sec: 3173.9, 60 sec: 2850.0, 300 sec: 2915.6). Total num frames: 805888. Throughput: 0: 2919.1. Samples: 798590. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:24:22,715][16994] Avg episode reward: [(0, '34.144')]
[2023-03-03 11:24:24,004][17031] Updated weights for policy 0, policy_version 790 (0.0009)
[2023-03-03 11:24:27,681][16994] Fps is (10 sec: 2764.9, 60 sec: 2867.5, 300 sec: 2907.4). Total num frames: 818176. Throughput: 0: 2878.1. Samples: 814775. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:24:27,682][16994] Avg episode reward: [(0, '33.307')]
[2023-03-03 11:24:27,916][17031] Updated weights for policy 0, policy_version 800 (0.0018)
[2023-03-03 11:24:31,677][17031] Updated weights for policy 0, policy_version 810 (0.0009)
[2023-03-03 11:24:32,683][16994] Fps is (10 sec: 2560.2, 60 sec: 2884.3, 300 sec: 2903.1). Total num frames: 831488. Throughput: 0: 2898.5. Samples: 830715. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:24:32,685][16994] Avg episode reward: [(0, '36.157')]
[2023-03-03 11:24:35,673][17031] Updated weights for policy 0, policy_version 820 (0.0008)
[2023-03-03 11:24:37,683][16994] Fps is (10 sec: 2457.2, 60 sec: 2816.0, 300 sec: 2891.9). Total num frames: 842752. Throughput: 0: 2909.6. Samples: 838873. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:24:37,686][16994] Avg episode reward: [(0, '36.837')]
[2023-03-03 11:24:39,860][17031] Updated weights for policy 0, policy_version 830 (0.0007)
[2023-03-03 11:24:42,684][16994] Fps is (10 sec: 2661.9, 60 sec: 2833.0, 300 sec: 2895.0). Total num frames: 858112. Throughput: 0: 2868.5. Samples: 853563. Policy #0 lag: (min: 0.0, avg: 0.9, max: 1.0)
[2023-03-03 11:24:42,686][16994] Avg episode reward: [(0, '38.594')]
[2023-03-03 11:24:43,352][17031] Updated weights for policy 0, policy_version 840 (0.0008)
[2023-03-03 11:24:46,685][17031] Updated weights for policy 0, policy_version 850 (0.0007)
[2023-03-03 11:24:47,684][16994] Fps is (10 sec: 2969.1, 60 sec: 2867.2, 300 sec: 2888.0). Total num frames: 872448. Throughput: 0: 2854.9. Samples: 871778. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:24:47,685][16994] Avg episode reward: [(0, '75.849')]
[2023-03-03 11:24:47,749][17030] Saving new best policy, reward=75.849!
[2023-03-03 11:24:49,993][17031] Updated weights for policy 0, policy_version 860 (0.0006)
[2023-03-03 11:24:52,682][16994] Fps is (10 sec: 2970.3, 60 sec: 2901.4, 300 sec: 2888.0). Total num frames: 887808. Throughput: 0: 2861.7. Samples: 881159. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:24:52,683][16994] Avg episode reward: [(0, '68.514')]
[2023-03-03 11:24:53,447][17031] Updated weights for policy 0, policy_version 870 (0.0007)
[2023-03-03 11:24:56,946][17031] Updated weights for policy 0, policy_version 880 (0.0007)
[2023-03-03 11:24:57,682][16994] Fps is (10 sec: 3072.8, 60 sec: 2918.4, 300 sec: 2888.0). Total num frames: 903168. Throughput: 0: 2869.7. Samples: 898967. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:24:57,684][16994] Avg episode reward: [(0, '62.574')]
[2023-03-03 11:25:00,524][17031] Updated weights for policy 0, policy_version 890 (0.0008)
[2023-03-03 11:25:02,684][16994] Fps is (10 sec: 2866.8, 60 sec: 2867.2, 300 sec: 2877.6). Total num frames: 916480. Throughput: 0: 2816.2. Samples: 915833. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:25:02,686][16994] Avg episode reward: [(0, '24.017')]
[2023-03-03 11:25:04,142][17031] Updated weights for policy 0, policy_version 900 (0.0008)
[2023-03-03 11:25:07,639][17031] Updated weights for policy 0, policy_version 910 (0.0008)
[2023-03-03 11:25:07,682][16994] Fps is (10 sec: 2867.2, 60 sec: 2884.4, 300 sec: 2881.1). Total num frames: 931840. Throughput: 0: 2794.2. Samples: 924326. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:25:07,683][16994] Avg episode reward: [(0, '28.655')]
[2023-03-03 11:25:10,859][17031] Updated weights for policy 0, policy_version 920 (0.0007)
[2023-03-03 11:25:12,682][16994] Fps is (10 sec: 3072.5, 60 sec: 2884.3, 300 sec: 2888.0). Total num frames: 947200. Throughput: 0: 2848.7. Samples: 942970. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:25:12,682][16994] Avg episode reward: [(0, '31.668')]
[2023-03-03 11:25:14,066][17031] Updated weights for policy 0, policy_version 930 (0.0006)
[2023-03-03 11:25:17,074][17031] Updated weights for policy 0, policy_version 940 (0.0007)
[2023-03-03 11:25:17,682][16994] Fps is (10 sec: 3174.1, 60 sec: 2884.2, 300 sec: 2895.0). Total num frames: 963584. Throughput: 0: 2933.8. Samples: 962737. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:25:17,684][16994] Avg episode reward: [(0, '32.671')]
[2023-03-03 11:25:20,145][17031] Updated weights for policy 0, policy_version 950 (0.0007)
[2023-03-03 11:25:22,684][16994] Fps is (10 sec: 3276.0, 60 sec: 2901.3, 300 sec: 2898.4). Total num frames: 979968. Throughput: 0: 2974.6. Samples: 972736. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:25:22,688][16994] Avg episode reward: [(0, '29.464')]
[2023-03-03 11:25:23,718][17031] Updated weights for policy 0, policy_version 960 (0.0010)
[2023-03-03 11:25:27,144][17031] Updated weights for policy 0, policy_version 970 (0.0008)
[2023-03-03 11:25:27,684][16994] Fps is (10 sec: 3072.0, 60 sec: 2935.4, 300 sec: 2891.5). Total num frames: 994304. Throughput: 0: 3039.9. Samples: 990355. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:25:27,691][16994] Avg episode reward: [(0, '26.950')]
[2023-03-03 11:25:30,482][17031] Updated weights for policy 0, policy_version 980 (0.0007)
[2023-03-03 11:25:32,684][16994] Fps is (10 sec: 2969.7, 60 sec: 2969.5, 300 sec: 2888.0). Total num frames: 1009664. Throughput: 0: 3038.2. Samples: 1008497. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:25:32,685][16994] Avg episode reward: [(0, '24.933')]
[2023-03-03 11:25:33,873][17031] Updated weights for policy 0, policy_version 990 (0.0008)
[2023-03-03 11:25:37,336][17031] Updated weights for policy 0, policy_version 1000 (0.0007)
[2023-03-03 11:25:37,683][16994] Fps is (10 sec: 3071.8, 60 sec: 3037.8, 300 sec: 2891.5). Total num frames: 1025024. Throughput: 0: 3025.2. Samples: 1017296. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:25:37,684][16994] Avg episode reward: [(0, '23.356')]
[2023-03-03 11:25:37,688][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001001_1025024.pth...
[2023-03-03 11:25:37,785][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000000318_325632.pth
[2023-03-03 11:25:40,543][17031] Updated weights for policy 0, policy_version 1010 (0.0007)
[2023-03-03 11:25:42,680][16994] Fps is (10 sec: 3073.2, 60 sec: 3038.1, 300 sec: 2895.0). Total num frames: 1040384. Throughput: 0: 3050.9. Samples: 1036254. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:25:42,681][16994] Avg episode reward: [(0, '21.580')]
[2023-03-03 11:25:43,708][17031] Updated weights for policy 0, policy_version 1020 (0.0007)
[2023-03-03 11:25:46,853][17031] Updated weights for policy 0, policy_version 1030 (0.0007)
[2023-03-03 11:25:47,681][16994] Fps is (10 sec: 3175.2, 60 sec: 3072.2, 300 sec: 2912.3). Total num frames: 1056768. Throughput: 0: 3106.1. Samples: 1055600. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:25:47,681][16994] Avg episode reward: [(0, '21.689')]
[2023-03-03 11:25:50,305][17031] Updated weights for policy 0, policy_version 1040 (0.0007)
[2023-03-03 11:25:52,683][16994] Fps is (10 sec: 3071.3, 60 sec: 3054.9, 300 sec: 2908.9). Total num frames: 1071104. Throughput: 0: 3115.0. Samples: 1064505. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:25:52,687][16994] Avg episode reward: [(0, '21.957')]
[2023-03-03 11:25:53,894][17031] Updated weights for policy 0, policy_version 1050 (0.0008)
[2023-03-03 11:25:57,684][16994] Fps is (10 sec: 2763.9, 60 sec: 3020.7, 300 sec: 2908.8). Total num frames: 1084416. Throughput: 0: 3062.0. Samples: 1080768. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:25:57,687][16994] Avg episode reward: [(0, '24.141')]
[2023-03-03 11:25:57,993][17031] Updated weights for policy 0, policy_version 1060 (0.0012)
[2023-03-03 11:26:01,546][17031] Updated weights for policy 0, policy_version 1070 (0.0007)
[2023-03-03 11:26:02,681][16994] Fps is (10 sec: 2765.2, 60 sec: 3038.0, 300 sec: 2908.9). Total num frames: 1098752. Throughput: 0: 2994.9. Samples: 1097505. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:02,682][16994] Avg episode reward: [(0, '25.576')]
[2023-03-03 11:26:04,773][17031] Updated weights for policy 0, policy_version 1080 (0.0006)
[2023-03-03 11:26:07,682][16994] Fps is (10 sec: 2970.1, 60 sec: 3037.8, 300 sec: 2908.9). Total num frames: 1114112. Throughput: 0: 2987.5. Samples: 1107167. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:26:07,683][16994] Avg episode reward: [(0, '25.947')]
[2023-03-03 11:26:08,026][17031] Updated weights for policy 0, policy_version 1090 (0.0007)
[2023-03-03 11:26:11,492][17031] Updated weights for policy 0, policy_version 1100 (0.0008)
[2023-03-03 11:26:12,682][16994] Fps is (10 sec: 2969.4, 60 sec: 3020.8, 300 sec: 2908.9). Total num frames: 1128448. Throughput: 0: 3001.8. Samples: 1125435. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:12,685][16994] Avg episode reward: [(0, '25.543')]
[2023-03-03 11:26:15,267][17031] Updated weights for policy 0, policy_version 1110 (0.0007)
[2023-03-03 11:26:17,685][16994] Fps is (10 sec: 2866.4, 60 sec: 2986.5, 300 sec: 2908.8). Total num frames: 1142784. Throughput: 0: 2980.6. Samples: 1142628. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:26:17,687][16994] Avg episode reward: [(0, '26.298')]
[2023-03-03 11:26:18,892][17031] Updated weights for policy 0, policy_version 1120 (0.0007)
[2023-03-03 11:26:22,356][17031] Updated weights for policy 0, policy_version 1130 (0.0007)
[2023-03-03 11:26:22,683][16994] Fps is (10 sec: 2969.3, 60 sec: 2969.7, 300 sec: 2915.8). Total num frames: 1158144. Throughput: 0: 2962.9. Samples: 1150626. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:26:22,685][16994] Avg episode reward: [(0, '26.261')]
[2023-03-03 11:26:25,919][17031] Updated weights for policy 0, policy_version 1140 (0.0007)
[2023-03-03 11:26:27,685][16994] Fps is (10 sec: 2969.7, 60 sec: 2969.5, 300 sec: 2922.7). Total num frames: 1172480. Throughput: 0: 2929.2. Samples: 1168079. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:27,693][16994] Avg episode reward: [(0, '25.393')]
[2023-03-03 11:26:29,464][17031] Updated weights for policy 0, policy_version 1150 (0.0008)
[2023-03-03 11:26:32,683][16994] Fps is (10 sec: 2764.7, 60 sec: 2935.5, 300 sec: 2919.3). Total num frames: 1185792. Throughput: 0: 2869.7. Samples: 1184742. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:32,685][16994] Avg episode reward: [(0, '25.573')]
[2023-03-03 11:26:33,225][17031] Updated weights for policy 0, policy_version 1160 (0.0008)
[2023-03-03 11:26:36,844][17031] Updated weights for policy 0, policy_version 1170 (0.0008)
[2023-03-03 11:26:37,682][16994] Fps is (10 sec: 2765.4, 60 sec: 2918.4, 300 sec: 2926.2). Total num frames: 1200128. Throughput: 0: 2852.0. Samples: 1192844. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:37,684][16994] Avg episode reward: [(0, '23.795')]
[2023-03-03 11:26:40,127][17031] Updated weights for policy 0, policy_version 1180 (0.0007)
[2023-03-03 11:26:42,683][16994] Fps is (10 sec: 2969.4, 60 sec: 2918.2, 300 sec: 2929.7). Total num frames: 1215488. Throughput: 0: 2908.0. Samples: 1211628. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:26:42,685][16994] Avg episode reward: [(0, '26.292')]
[2023-03-03 11:26:43,477][17031] Updated weights for policy 0, policy_version 1190 (0.0008)
[2023-03-03 11:26:46,966][17031] Updated weights for policy 0, policy_version 1200 (0.0007)
[2023-03-03 11:26:47,683][16994] Fps is (10 sec: 2969.4, 60 sec: 2884.1, 300 sec: 2929.7). Total num frames: 1229824. Throughput: 0: 2930.1. Samples: 1229364. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:26:47,686][16994] Avg episode reward: [(0, '26.180')]
[2023-03-03 11:26:50,495][17031] Updated weights for policy 0, policy_version 1210 (0.0007)
[2023-03-03 11:26:52,682][16994] Fps is (10 sec: 2970.0, 60 sec: 2901.3, 300 sec: 2929.7). Total num frames: 1245184. Throughput: 0: 2914.4. Samples: 1238314. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:26:52,684][16994] Avg episode reward: [(0, '27.147')]
[2023-03-03 11:26:54,010][17031] Updated weights for policy 0, policy_version 1220 (0.0007)
[2023-03-03 11:26:57,685][16994] Fps is (10 sec: 2866.7, 60 sec: 2901.3, 300 sec: 2922.7). Total num frames: 1258496. Throughput: 0: 2877.7. Samples: 1254939. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:26:57,687][16994] Avg episode reward: [(0, '26.511')]
[2023-03-03 11:26:57,846][17031] Updated weights for policy 0, policy_version 1230 (0.0008)
[2023-03-03 11:27:01,517][17031] Updated weights for policy 0, policy_version 1240 (0.0008)
[2023-03-03 11:27:02,684][16994] Fps is (10 sec: 2764.4, 60 sec: 2901.2, 300 sec: 2919.3). Total num frames: 1272832. Throughput: 0: 2865.9. Samples: 1271592. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:27:02,685][16994] Avg episode reward: [(0, '26.618')]
[2023-03-03 11:27:05,042][17031] Updated weights for policy 0, policy_version 1250 (0.0008)
[2023-03-03 11:27:07,683][16994] Fps is (10 sec: 2867.8, 60 sec: 2884.2, 300 sec: 2926.2). Total num frames: 1287168. Throughput: 0: 2883.1. Samples: 1280366. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:07,684][16994] Avg episode reward: [(0, '26.821')]
[2023-03-03 11:27:08,349][17031] Updated weights for policy 0, policy_version 1260 (0.0006)
[2023-03-03 11:27:11,975][17031] Updated weights for policy 0, policy_version 1270 (0.0007)
[2023-03-03 11:27:12,682][16994] Fps is (10 sec: 2970.2, 60 sec: 2901.3, 300 sec: 2926.2). Total num frames: 1302528. Throughput: 0: 2890.0. Samples: 1298124. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:12,683][16994] Avg episode reward: [(0, '27.536')]
[2023-03-03 11:27:15,328][17031] Updated weights for policy 0, policy_version 1280 (0.0007)
[2023-03-03 11:27:17,684][16994] Fps is (10 sec: 3071.7, 60 sec: 2918.5, 300 sec: 2926.2). Total num frames: 1317888. Throughput: 0: 2929.7. Samples: 1316578. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:27:17,685][16994] Avg episode reward: [(0, '28.745')]
[2023-03-03 11:27:19,199][17031] Updated weights for policy 0, policy_version 1290 (0.0007)
[2023-03-03 11:27:22,683][16994] Fps is (10 sec: 2764.7, 60 sec: 2867.2, 300 sec: 2922.8). Total num frames: 1330176. Throughput: 0: 2904.9. Samples: 1323565. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:22,686][16994] Avg episode reward: [(0, '28.652')]
[2023-03-03 11:27:22,860][17031] Updated weights for policy 0, policy_version 1300 (0.0007)
[2023-03-03 11:27:26,179][17031] Updated weights for policy 0, policy_version 1310 (0.0008)
[2023-03-03 11:27:27,684][16994] Fps is (10 sec: 2764.6, 60 sec: 2884.3, 300 sec: 2919.3). Total num frames: 1345536. Throughput: 0: 2880.4. Samples: 1341249. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:27,685][16994] Avg episode reward: [(0, '28.455')]
[2023-03-03 11:27:29,291][17031] Updated weights for policy 0, policy_version 1320 (0.0007)
[2023-03-03 11:27:32,384][17031] Updated weights for policy 0, policy_version 1330 (0.0007)
[2023-03-03 11:27:32,680][16994] Fps is (10 sec: 3277.5, 60 sec: 2952.7, 300 sec: 2929.7). Total num frames: 1362944. Throughput: 0: 2927.7. Samples: 1361100. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:27:32,682][16994] Avg episode reward: [(0, '27.391')]
[2023-03-03 11:27:35,647][17031] Updated weights for policy 0, policy_version 1340 (0.0007)
[2023-03-03 11:27:37,684][16994] Fps is (10 sec: 3174.7, 60 sec: 2952.5, 300 sec: 2929.7). Total num frames: 1377280. Throughput: 0: 2943.8. Samples: 1370790. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:37,685][16994] Avg episode reward: [(0, '28.191')]
[2023-03-03 11:27:37,843][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001346_1378304.pth...
[2023-03-03 11:27:37,943][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000000659_674816.pth
[2023-03-03 11:27:39,245][17031] Updated weights for policy 0, policy_version 1350 (0.0007)
[2023-03-03 11:27:42,677][17031] Updated weights for policy 0, policy_version 1360 (0.0006)
[2023-03-03 11:27:42,681][16994] Fps is (10 sec: 2969.4, 60 sec: 2952.7, 300 sec: 2933.2). Total num frames: 1392640. Throughput: 0: 2972.9. Samples: 1388707. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:27:42,682][16994] Avg episode reward: [(0, '27.458')]
[2023-03-03 11:27:45,708][17031] Updated weights for policy 0, policy_version 1370 (0.0006)
[2023-03-03 11:27:47,682][16994] Fps is (10 sec: 3072.6, 60 sec: 2969.7, 300 sec: 2933.2). Total num frames: 1408000. Throughput: 0: 3013.3. Samples: 1407182. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:27:47,682][16994] Avg episode reward: [(0, '31.141')]
[2023-03-03 11:27:49,205][17031] Updated weights for policy 0, policy_version 1380 (0.0007)
[2023-03-03 11:27:52,680][16994] Fps is (10 sec: 2969.8, 60 sec: 2952.6, 300 sec: 2926.2). Total num frames: 1422336. Throughput: 0: 3011.7. Samples: 1415884. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:27:52,681][16994] Avg episode reward: [(0, '30.809')]
[2023-03-03 11:27:52,696][17031] Updated weights for policy 0, policy_version 1390 (0.0008)
[2023-03-03 11:27:56,033][17031] Updated weights for policy 0, policy_version 1400 (0.0007)
[2023-03-03 11:27:57,681][16994] Fps is (10 sec: 2969.6, 60 sec: 2986.8, 300 sec: 2922.8). Total num frames: 1437696. Throughput: 0: 3018.2. Samples: 1433942. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:27:57,683][16994] Avg episode reward: [(0, '33.015')]
[2023-03-03 11:27:59,854][17031] Updated weights for policy 0, policy_version 1410 (0.0008)
[2023-03-03 11:28:02,684][16994] Fps is (10 sec: 2866.1, 60 sec: 2969.6, 300 sec: 2915.8). Total num frames: 1451008. Throughput: 0: 2975.3. Samples: 1450471. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:28:02,687][16994] Avg episode reward: [(0, '38.885')]
[2023-03-03 11:28:03,666][17031] Updated weights for policy 0, policy_version 1420 (0.0008)
[2023-03-03 11:28:07,413][17031] Updated weights for policy 0, policy_version 1430 (0.0008)
[2023-03-03 11:28:07,690][16994] Fps is (10 sec: 2660.7, 60 sec: 2952.3, 300 sec: 2912.3). Total num frames: 1464320. Throughput: 0: 2989.5. Samples: 1458109. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:28:07,701][16994] Avg episode reward: [(0, '50.264')]
[2023-03-03 11:28:10,743][17031] Updated weights for policy 0, policy_version 1440 (0.0007)
[2023-03-03 11:28:12,681][16994] Fps is (10 sec: 2868.2, 60 sec: 2952.6, 300 sec: 2916.0). Total num frames: 1479680. Throughput: 0: 2992.3. Samples: 1475893. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:28:12,682][16994] Avg episode reward: [(0, '50.518')]
[2023-03-03 11:28:14,239][17031] Updated weights for policy 0, policy_version 1450 (0.0007)
[2023-03-03 11:28:17,486][17031] Updated weights for policy 0, policy_version 1460 (0.0008)
[2023-03-03 11:28:17,684][16994] Fps is (10 sec: 3073.3, 60 sec: 2952.5, 300 sec: 2915.8). Total num frames: 1495040. Throughput: 0: 2948.7. Samples: 1493800. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:17,685][16994] Avg episode reward: [(0, '44.278')]
[2023-03-03 11:28:21,119][17031] Updated weights for policy 0, policy_version 1470 (0.0009)
[2023-03-03 11:28:22,685][16994] Fps is (10 sec: 2968.3, 60 sec: 2986.5, 300 sec: 2926.2). Total num frames: 1509376. Throughput: 0: 2925.1. Samples: 1502423. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:22,688][16994] Avg episode reward: [(0, '41.011')]
[2023-03-03 11:28:24,637][17031] Updated weights for policy 0, policy_version 1480 (0.0008)
[2023-03-03 11:28:27,684][16994] Fps is (10 sec: 2764.6, 60 sec: 2952.5, 300 sec: 2929.7). Total num frames: 1522688. Throughput: 0: 2924.3. Samples: 1520311. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:27,686][16994] Avg episode reward: [(0, '41.027')]
[2023-03-03 11:28:28,430][17031] Updated weights for policy 0, policy_version 1490 (0.0007)
[2023-03-03 11:28:32,298][17031] Updated weights for policy 0, policy_version 1500 (0.0008)
[2023-03-03 11:28:32,681][16994] Fps is (10 sec: 2766.0, 60 sec: 2901.3, 300 sec: 2926.2). Total num frames: 1537024. Throughput: 0: 2851.7. Samples: 1535505. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:32,682][16994] Avg episode reward: [(0, '40.781')]
[2023-03-03 11:28:36,012][17031] Updated weights for policy 0, policy_version 1510 (0.0007)
[2023-03-03 11:28:37,681][16994] Fps is (10 sec: 2868.1, 60 sec: 2901.4, 300 sec: 2926.2). Total num frames: 1551360. Throughput: 0: 2844.2. Samples: 1543876. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:37,682][16994] Avg episode reward: [(0, '42.893')]
[2023-03-03 11:28:39,190][17031] Updated weights for policy 0, policy_version 1520 (0.0006)
[2023-03-03 11:28:42,312][17031] Updated weights for policy 0, policy_version 1530 (0.0007)
[2023-03-03 11:28:42,683][16994] Fps is (10 sec: 2969.0, 60 sec: 2901.2, 300 sec: 2936.6). Total num frames: 1566720. Throughput: 0: 2872.9. Samples: 1563228. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:42,684][16994] Avg episode reward: [(0, '35.819')]
[2023-03-03 11:28:45,701][17031] Updated weights for policy 0, policy_version 1540 (0.0007)
[2023-03-03 11:28:47,684][16994] Fps is (10 sec: 3173.6, 60 sec: 2918.3, 300 sec: 2947.0). Total num frames: 1583104. Throughput: 0: 2914.7. Samples: 1581633. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:47,685][16994] Avg episode reward: [(0, '32.136')]
[2023-03-03 11:28:48,839][17031] Updated weights for policy 0, policy_version 1550 (0.0006)
[2023-03-03 11:28:52,614][17031] Updated weights for policy 0, policy_version 1560 (0.0008)
[2023-03-03 11:28:52,684][16994] Fps is (10 sec: 3071.8, 60 sec: 2918.2, 300 sec: 2947.0). Total num frames: 1597440. Throughput: 0: 2954.4. Samples: 1591045. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:52,687][16994] Avg episode reward: [(0, '26.815')]
[2023-03-03 11:28:55,831][17031] Updated weights for policy 0, policy_version 1570 (0.0007)
[2023-03-03 11:28:57,682][16994] Fps is (10 sec: 2867.9, 60 sec: 2901.3, 300 sec: 2940.1). Total num frames: 1611776. Throughput: 0: 2951.6. Samples: 1608718. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:28:57,684][16994] Avg episode reward: [(0, '27.026')]
[2023-03-03 11:29:00,070][17031] Updated weights for policy 0, policy_version 1580 (0.0008)
[2023-03-03 11:29:02,683][16994] Fps is (10 sec: 2764.8, 60 sec: 2901.4, 300 sec: 2936.6). Total num frames: 1625088. Throughput: 0: 2893.0. Samples: 1623983. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:29:02,685][16994] Avg episode reward: [(0, '27.992')]
[2023-03-03 11:29:03,608][17031] Updated weights for policy 0, policy_version 1590 (0.0008)
[2023-03-03 11:29:06,891][17031] Updated weights for policy 0, policy_version 1600 (0.0008)
[2023-03-03 11:29:07,684][16994] Fps is (10 sec: 2866.6, 60 sec: 2935.7, 300 sec: 2936.6). Total num frames: 1640448. Throughput: 0: 2904.8. Samples: 1633136. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:29:07,685][16994] Avg episode reward: [(0, '27.508')]
[2023-03-03 11:29:10,317][17031] Updated weights for policy 0, policy_version 1610 (0.0006)
[2023-03-03 11:29:12,684][16994] Fps is (10 sec: 3071.9, 60 sec: 2935.3, 300 sec: 2933.1). Total num frames: 1655808. Throughput: 0: 2915.0. Samples: 1651484. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:29:12,686][16994] Avg episode reward: [(0, '28.500')]
[2023-03-03 11:29:13,581][17031] Updated weights for policy 0, policy_version 1620 (0.0007)
[2023-03-03 11:29:17,015][17031] Updated weights for policy 0, policy_version 1630 (0.0007)
[2023-03-03 11:29:17,683][16994] Fps is (10 sec: 2969.8, 60 sec: 2918.4, 300 sec: 2929.7). Total num frames: 1670144. Throughput: 0: 2981.3. Samples: 1669669. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:29:17,684][16994] Avg episode reward: [(0, '26.629')]
[2023-03-03 11:29:21,099][17031] Updated weights for policy 0, policy_version 1640 (0.0009)
[2023-03-03 11:29:22,683][16994] Fps is (10 sec: 2764.9, 60 sec: 2901.4, 300 sec: 2933.1). Total num frames: 1683456. Throughput: 0: 2954.3. Samples: 1676827. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:29:22,686][16994] Avg episode reward: [(0, '25.562')]
[2023-03-03 11:29:24,513][17031] Updated weights for policy 0, policy_version 1650 (0.0007)
[2023-03-03 11:29:27,683][16994] Fps is (10 sec: 2867.3, 60 sec: 2935.5, 300 sec: 2940.1). Total num frames: 1698816. Throughput: 0: 2919.1. Samples: 1694587. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:29:27,683][16994] Avg episode reward: [(0, '24.832')]
[2023-03-03 11:29:27,787][17031] Updated weights for policy 0, policy_version 1660 (0.0007)
[2023-03-03 11:29:31,098][17031] Updated weights for policy 0, policy_version 1670 (0.0007)
[2023-03-03 11:29:32,682][16994] Fps is (10 sec: 3174.6, 60 sec: 2969.5, 300 sec: 2957.5). Total num frames: 1715200. Throughput: 0: 2933.9. Samples: 1713654. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:29:32,684][16994] Avg episode reward: [(0, '24.230')]
[2023-03-03 11:29:34,319][17031] Updated weights for policy 0, policy_version 1680 (0.0006)
[2023-03-03 11:29:37,551][17031] Updated weights for policy 0, policy_version 1690 (0.0007)
[2023-03-03 11:29:37,682][16994] Fps is (10 sec: 3174.7, 60 sec: 2986.6, 300 sec: 2957.5). Total num frames: 1730560. Throughput: 0: 2934.9. Samples: 1723113. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:29:37,682][16994] Avg episode reward: [(0, '22.712')]
[2023-03-03 11:29:37,848][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001691_1731584.pth...
[2023-03-03 11:29:37,925][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001001_1025024.pth
[2023-03-03 11:29:40,724][17031] Updated weights for policy 0, policy_version 1700 (0.0006)
[2023-03-03 11:29:42,681][16994] Fps is (10 sec: 3174.9, 60 sec: 3003.8, 300 sec: 2964.4). Total num frames: 1746944. Throughput: 0: 2964.8. Samples: 1742132. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:29:42,681][16994] Avg episode reward: [(0, '33.891')]
[2023-03-03 11:29:43,965][17031] Updated weights for policy 0, policy_version 1710 (0.0007)
[2023-03-03 11:29:47,154][17031] Updated weights for policy 0, policy_version 1720 (0.0007)
[2023-03-03 11:29:47,684][16994] Fps is (10 sec: 3173.7, 60 sec: 2986.7, 300 sec: 2964.4). Total num frames: 1762304. Throughput: 0: 3049.4. Samples: 1761207. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:29:47,686][16994] Avg episode reward: [(0, '35.401')]
[2023-03-03 11:29:50,291][17031] Updated weights for policy 0, policy_version 1730 (0.0007)
[2023-03-03 11:29:52,684][16994] Fps is (10 sec: 3071.1, 60 sec: 3003.7, 300 sec: 2964.4). Total num frames: 1777664. Throughput: 0: 3063.1. Samples: 1770973. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:29:52,685][16994] Avg episode reward: [(0, '43.094')]
[2023-03-03 11:29:53,638][17031] Updated weights for policy 0, policy_version 1740 (0.0007)
[2023-03-03 11:29:56,850][17031] Updated weights for policy 0, policy_version 1750 (0.0007)
[2023-03-03 11:29:57,683][16994] Fps is (10 sec: 3174.6, 60 sec: 3037.8, 300 sec: 2974.8). Total num frames: 1794048. Throughput: 0: 3070.8. Samples: 1789670. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:29:57,684][16994] Avg episode reward: [(0, '28.493')]
[2023-03-03 11:30:00,060][17031] Updated weights for policy 0, policy_version 1760 (0.0007)
[2023-03-03 11:30:02,684][16994] Fps is (10 sec: 3276.7, 60 sec: 3089.0, 300 sec: 2978.3). Total num frames: 1810432. Throughput: 0: 3095.1. Samples: 1808953. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:02,686][16994] Avg episode reward: [(0, '29.215')]
[2023-03-03 11:30:03,293][17031] Updated weights for policy 0, policy_version 1770 (0.0006)
[2023-03-03 11:30:07,087][17031] Updated weights for policy 0, policy_version 1780 (0.0008)
[2023-03-03 11:30:07,683][16994] Fps is (10 sec: 2969.6, 60 sec: 3055.0, 300 sec: 2971.3). Total num frames: 1823744. Throughput: 0: 3130.3. Samples: 1817691. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:30:07,685][16994] Avg episode reward: [(0, '23.380')]
[2023-03-03 11:30:10,616][17031] Updated weights for policy 0, policy_version 1790 (0.0008)
[2023-03-03 11:30:12,684][16994] Fps is (10 sec: 2867.1, 60 sec: 3054.9, 300 sec: 2967.8). Total num frames: 1839104. Throughput: 0: 3110.6. Samples: 1834569. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:12,686][16994] Avg episode reward: [(0, '30.477')]
[2023-03-03 11:30:14,127][17031] Updated weights for policy 0, policy_version 1800 (0.0009)
[2023-03-03 11:30:17,487][17031] Updated weights for policy 0, policy_version 1810 (0.0007)
[2023-03-03 11:30:17,685][16994] Fps is (10 sec: 2969.2, 60 sec: 3054.8, 300 sec: 2960.9). Total num frames: 1853440. Throughput: 0: 3078.5. Samples: 1852192. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:30:17,686][16994] Avg episode reward: [(0, '33.906')]
[2023-03-03 11:30:20,963][17031] Updated weights for policy 0, policy_version 1820 (0.0008)
[2023-03-03 11:30:22,684][16994] Fps is (10 sec: 2867.3, 60 sec: 3071.9, 300 sec: 2960.9). Total num frames: 1867776. Throughput: 0: 3076.0. Samples: 1861539. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:30:22,685][16994] Avg episode reward: [(0, '38.144')]
[2023-03-03 11:30:24,624][17031] Updated weights for policy 0, policy_version 1830 (0.0008)
[2023-03-03 11:30:27,681][16994] Fps is (10 sec: 2868.1, 60 sec: 3055.0, 300 sec: 2957.5). Total num frames: 1882112. Throughput: 0: 3028.8. Samples: 1878432. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:27,682][16994] Avg episode reward: [(0, '32.024')]
[2023-03-03 11:30:28,129][17031] Updated weights for policy 0, policy_version 1840 (0.0007)
[2023-03-03 11:30:31,670][17031] Updated weights for policy 0, policy_version 1850 (0.0010)
[2023-03-03 11:30:32,683][16994] Fps is (10 sec: 2867.7, 60 sec: 3020.8, 300 sec: 2954.0). Total num frames: 1896448. Throughput: 0: 2992.6. Samples: 1895870. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:32,685][16994] Avg episode reward: [(0, '29.187')]
[2023-03-03 11:30:35,411][17031] Updated weights for policy 0, policy_version 1860 (0.0008)
[2023-03-03 11:30:37,692][16994] Fps is (10 sec: 2864.1, 60 sec: 3003.2, 300 sec: 2950.4). Total num frames: 1910784. Throughput: 0: 2956.7. Samples: 1904050. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:37,696][16994] Avg episode reward: [(0, '21.782')]
[2023-03-03 11:30:39,215][17031] Updated weights for policy 0, policy_version 1870 (0.0008)
[2023-03-03 11:30:42,438][17031] Updated weights for policy 0, policy_version 1880 (0.0006)
[2023-03-03 11:30:42,682][16994] Fps is (10 sec: 2867.2, 60 sec: 2969.5, 300 sec: 2943.5). Total num frames: 1925120. Throughput: 0: 2917.4. Samples: 1920950. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:42,683][16994] Avg episode reward: [(0, '21.178')]
[2023-03-03 11:30:45,660][17031] Updated weights for policy 0, policy_version 1890 (0.0007)
[2023-03-03 11:30:47,683][16994] Fps is (10 sec: 2972.3, 60 sec: 2969.6, 300 sec: 2947.0). Total num frames: 1940480. Throughput: 0: 2901.5. Samples: 1939519. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:47,684][16994] Avg episode reward: [(0, '16.938')]
[2023-03-03 11:30:49,156][17031] Updated weights for policy 0, policy_version 1900 (0.0007)
[2023-03-03 11:30:52,683][16994] Fps is (10 sec: 2969.4, 60 sec: 2952.6, 300 sec: 2950.5). Total num frames: 1954816. Throughput: 0: 2905.4. Samples: 1948432. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:52,684][16994] Avg episode reward: [(0, '18.879')]
[2023-03-03 11:30:52,687][17031] Updated weights for policy 0, policy_version 1910 (0.0007)
[2023-03-03 11:30:56,121][17031] Updated weights for policy 0, policy_version 1920 (0.0007)
[2023-03-03 11:30:57,682][16994] Fps is (10 sec: 2969.9, 60 sec: 2935.5, 300 sec: 2954.0). Total num frames: 1970176. Throughput: 0: 2924.8. Samples: 1966178. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:30:57,683][16994] Avg episode reward: [(0, '21.724')]
[2023-03-03 11:30:59,453][17031] Updated weights for policy 0, policy_version 1930 (0.0007)
[2023-03-03 11:31:02,681][16994] Fps is (10 sec: 3072.7, 60 sec: 2918.5, 300 sec: 2954.0). Total num frames: 1985536. Throughput: 0: 2934.2. Samples: 1984222. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:31:02,682][16994] Avg episode reward: [(0, '33.330')]
[2023-03-03 11:31:02,987][17031] Updated weights for policy 0, policy_version 1940 (0.0007)
[2023-03-03 11:31:06,354][17031] Updated weights for policy 0, policy_version 1950 (0.0007)
[2023-03-03 11:31:07,682][16994] Fps is (10 sec: 3072.0, 60 sec: 2952.6, 300 sec: 2957.4). Total num frames: 2000896. Throughput: 0: 2922.2. Samples: 1993031. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:07,683][16994] Avg episode reward: [(0, '36.803')]
[2023-03-03 11:31:09,642][17031] Updated weights for policy 0, policy_version 1960 (0.0007)
[2023-03-03 11:31:12,682][16994] Fps is (10 sec: 3071.6, 60 sec: 2952.6, 300 sec: 2960.9). Total num frames: 2016256. Throughput: 0: 2957.6. Samples: 2011526. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:31:12,683][16994] Avg episode reward: [(0, '39.234')]
[2023-03-03 11:31:13,041][17031] Updated weights for policy 0, policy_version 1970 (0.0007)
[2023-03-03 11:31:16,297][17031] Updated weights for policy 0, policy_version 1980 (0.0007)
[2023-03-03 11:31:17,683][16994] Fps is (10 sec: 2969.3, 60 sec: 2952.6, 300 sec: 2957.4). Total num frames: 2030592. Throughput: 0: 2985.1. Samples: 2030203. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:17,684][16994] Avg episode reward: [(0, '29.976')]
[2023-03-03 11:31:19,785][17031] Updated weights for policy 0, policy_version 1990 (0.0007)
[2023-03-03 11:31:22,682][16994] Fps is (10 sec: 2969.6, 60 sec: 2969.7, 300 sec: 2960.9). Total num frames: 2045952. Throughput: 0: 2998.3. Samples: 2038945. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:22,685][16994] Avg episode reward: [(0, '29.448')]
[2023-03-03 11:31:23,168][17031] Updated weights for policy 0, policy_version 2000 (0.0008)
[2023-03-03 11:31:26,555][17031] Updated weights for policy 0, policy_version 2010 (0.0008)
[2023-03-03 11:31:27,684][16994] Fps is (10 sec: 3071.7, 60 sec: 2986.5, 300 sec: 2967.9). Total num frames: 2061312. Throughput: 0: 3024.7. Samples: 2057067. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:31:27,685][16994] Avg episode reward: [(0, '27.317')]
[2023-03-03 11:31:29,915][17031] Updated weights for policy 0, policy_version 2020 (0.0007)
[2023-03-03 11:31:32,683][16994] Fps is (10 sec: 3071.8, 60 sec: 3003.7, 300 sec: 2971.3). Total num frames: 2076672. Throughput: 0: 3021.4. Samples: 2075484. Policy #0 lag: (min: 0.0, avg: 0.9, max: 1.0)
[2023-03-03 11:31:32,683][16994] Avg episode reward: [(0, '24.669')]
[2023-03-03 11:31:33,231][17031] Updated weights for policy 0, policy_version 2030 (0.0014)
[2023-03-03 11:31:36,495][17031] Updated weights for policy 0, policy_version 2040 (0.0007)
[2023-03-03 11:31:37,684][16994] Fps is (10 sec: 3072.0, 60 sec: 3021.2, 300 sec: 2971.3). Total num frames: 2092032. Throughput: 0: 3030.1. Samples: 2084789. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:37,685][16994] Avg episode reward: [(0, '22.335')]
[2023-03-03 11:31:37,831][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002044_2093056.pth...
[2023-03-03 11:31:37,915][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001346_1378304.pth
[2023-03-03 11:31:39,691][17031] Updated weights for policy 0, policy_version 2050 (0.0007)
[2023-03-03 11:31:42,684][16994] Fps is (10 sec: 3174.2, 60 sec: 3054.9, 300 sec: 2978.3). Total num frames: 2108416. Throughput: 0: 3056.5. Samples: 2103725. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:42,684][16994] Avg episode reward: [(0, '21.471')]
[2023-03-03 11:31:42,939][17031] Updated weights for policy 0, policy_version 2060 (0.0006)
[2023-03-03 11:31:46,491][17031] Updated weights for policy 0, policy_version 2070 (0.0007)
[2023-03-03 11:31:47,683][16994] Fps is (10 sec: 3072.2, 60 sec: 3037.9, 300 sec: 2974.8). Total num frames: 2122752. Throughput: 0: 3053.3. Samples: 2121627. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:31:47,684][16994] Avg episode reward: [(0, '21.633')]
[2023-03-03 11:31:50,155][17031] Updated weights for policy 0, policy_version 2080 (0.0008)
[2023-03-03 11:31:52,684][16994] Fps is (10 sec: 2764.7, 60 sec: 3020.8, 300 sec: 2974.8). Total num frames: 2136064. Throughput: 0: 3041.7. Samples: 2129916. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:31:52,686][16994] Avg episode reward: [(0, '26.563')]
[2023-03-03 11:31:53,777][17031] Updated weights for policy 0, policy_version 2090 (0.0007)
[2023-03-03 11:31:57,684][16994] Fps is (10 sec: 2662.2, 60 sec: 2986.6, 300 sec: 2971.3). Total num frames: 2149376. Throughput: 0: 3005.1. Samples: 2146759. Policy #0 lag: (min: 0.0, avg: 0.6, max: 1.0)
[2023-03-03 11:31:57,685][16994] Avg episode reward: [(0, '27.451')]
[2023-03-03 11:31:57,828][17031] Updated weights for policy 0, policy_version 2100 (0.0008)
[2023-03-03 11:32:01,430][17031] Updated weights for policy 0, policy_version 2110 (0.0012)
[2023-03-03 11:32:02,685][16994] Fps is (10 sec: 2764.5, 60 sec: 2969.4, 300 sec: 2971.3). Total num frames: 2163712. Throughput: 0: 2939.2. Samples: 2162473. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:32:02,687][16994] Avg episode reward: [(0, '26.785')]
[2023-03-03 11:32:04,804][17031] Updated weights for policy 0, policy_version 2120 (0.0007)
[2023-03-03 11:32:07,681][16994] Fps is (10 sec: 3073.0, 60 sec: 2986.7, 300 sec: 2974.8). Total num frames: 2180096. Throughput: 0: 2953.5. Samples: 2171847. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:07,681][16994] Avg episode reward: [(0, '24.261')]
[2023-03-03 11:32:07,947][17031] Updated weights for policy 0, policy_version 2130 (0.0007)
[2023-03-03 11:32:11,117][17031] Updated weights for policy 0, policy_version 2140 (0.0007)
[2023-03-03 11:32:12,684][16994] Fps is (10 sec: 3277.3, 60 sec: 3003.7, 300 sec: 2978.3). Total num frames: 2196480. Throughput: 0: 2986.6. Samples: 2191461. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:32:12,684][16994] Avg episode reward: [(0, '20.585')]
[2023-03-03 11:32:14,130][17031] Updated weights for policy 0, policy_version 2150 (0.0007)
[2023-03-03 11:32:17,187][17031] Updated weights for policy 0, policy_version 2160 (0.0007)
[2023-03-03 11:32:17,685][16994] Fps is (10 sec: 3275.5, 60 sec: 3037.8, 300 sec: 2992.1). Total num frames: 2212864. Throughput: 0: 3029.0. Samples: 2211796. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:17,685][16994] Avg episode reward: [(0, '23.611')]
[2023-03-03 11:32:20,395][17031] Updated weights for policy 0, policy_version 2170 (0.0007)
[2023-03-03 11:32:22,684][16994] Fps is (10 sec: 3276.8, 60 sec: 3054.9, 300 sec: 2995.6). Total num frames: 2229248. Throughput: 0: 3030.3. Samples: 2221153. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:32:22,685][16994] Avg episode reward: [(0, '28.913')]
[2023-03-03 11:32:23,491][17031] Updated weights for policy 0, policy_version 2180 (0.0007)
[2023-03-03 11:32:26,598][17031] Updated weights for policy 0, policy_version 2190 (0.0007)
[2023-03-03 11:32:27,681][16994] Fps is (10 sec: 3278.0, 60 sec: 3072.2, 300 sec: 2992.2). Total num frames: 2245632. Throughput: 0: 3059.6. Samples: 2241401. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:32:27,681][16994] Avg episode reward: [(0, '36.480')]
[2023-03-03 11:32:29,700][17031] Updated weights for policy 0, policy_version 2200 (0.0007)
[2023-03-03 11:32:32,684][16994] Fps is (10 sec: 3276.7, 60 sec: 3089.0, 300 sec: 2999.1). Total num frames: 2262016. Throughput: 0: 3097.7. Samples: 2261024. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:32,685][16994] Avg episode reward: [(0, '34.381')]
[2023-03-03 11:32:32,704][17031] Updated weights for policy 0, policy_version 2210 (0.0006)
[2023-03-03 11:32:35,869][17031] Updated weights for policy 0, policy_version 2220 (0.0007)
[2023-03-03 11:32:37,684][16994] Fps is (10 sec: 3275.8, 60 sec: 3106.1, 300 sec: 3002.5). Total num frames: 2278400. Throughput: 0: 3130.5. Samples: 2270789. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:32:37,685][16994] Avg episode reward: [(0, '30.567')]
[2023-03-03 11:32:38,857][17031] Updated weights for policy 0, policy_version 2230 (0.0006)
[2023-03-03 11:32:41,867][17031] Updated weights for policy 0, policy_version 2240 (0.0006)
[2023-03-03 11:32:42,682][16994] Fps is (10 sec: 3379.8, 60 sec: 3123.3, 300 sec: 3009.5). Total num frames: 2295808. Throughput: 0: 3209.0. Samples: 2291160. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:42,683][16994] Avg episode reward: [(0, '35.577')]
[2023-03-03 11:32:44,906][17031] Updated weights for policy 0, policy_version 2250 (0.0006)
[2023-03-03 11:32:47,680][16994] Fps is (10 sec: 3483.0, 60 sec: 3174.6, 300 sec: 3019.9). Total num frames: 2313216. Throughput: 0: 3315.6. Samples: 2311658. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:47,681][16994] Avg episode reward: [(0, '42.404')]
[2023-03-03 11:32:47,952][17031] Updated weights for policy 0, policy_version 2260 (0.0006)
[2023-03-03 11:32:50,958][17031] Updated weights for policy 0, policy_version 2270 (0.0006)
[2023-03-03 11:32:52,683][16994] Fps is (10 sec: 3378.8, 60 sec: 3225.6, 300 sec: 3023.4). Total num frames: 2329600. Throughput: 0: 3334.5. Samples: 2321906. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:32:52,685][16994] Avg episode reward: [(0, '44.664')]
[2023-03-03 11:32:53,952][17031] Updated weights for policy 0, policy_version 2280 (0.0006)
[2023-03-03 11:32:56,997][17031] Updated weights for policy 0, policy_version 2290 (0.0006)
[2023-03-03 11:32:57,682][16994] Fps is (10 sec: 3378.6, 60 sec: 3294.0, 300 sec: 3037.3). Total num frames: 2347008. Throughput: 0: 3342.8. Samples: 2341882. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:32:57,683][16994] Avg episode reward: [(0, '52.804')]
[2023-03-03 11:33:00,038][17031] Updated weights for policy 0, policy_version 2300 (0.0007)
[2023-03-03 11:33:02,684][16994] Fps is (10 sec: 3379.1, 60 sec: 3328.1, 300 sec: 3047.7). Total num frames: 2363392. Throughput: 0: 3341.4. Samples: 2362157. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:33:02,684][16994] Avg episode reward: [(0, '47.901')]
[2023-03-03 11:33:03,050][17031] Updated weights for policy 0, policy_version 2310 (0.0006)
[2023-03-03 11:33:06,115][17031] Updated weights for policy 0, policy_version 2320 (0.0007)
[2023-03-03 11:33:07,684][16994] Fps is (10 sec: 3378.6, 60 sec: 3344.9, 300 sec: 3054.6). Total num frames: 2380800. Throughput: 0: 3359.9. Samples: 2372347. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:33:07,684][16994] Avg episode reward: [(0, '45.054')]
[2023-03-03 11:33:09,214][17031] Updated weights for policy 0, policy_version 2330 (0.0006)
[2023-03-03 11:33:12,214][17031] Updated weights for policy 0, policy_version 2340 (0.0006)
[2023-03-03 11:33:12,681][16994] Fps is (10 sec: 3380.2, 60 sec: 3345.2, 300 sec: 3058.1). Total num frames: 2397184. Throughput: 0: 3361.8. Samples: 2392682. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:33:12,681][16994] Avg episode reward: [(0, '26.640')]
[2023-03-03 11:33:15,310][17031] Updated weights for policy 0, policy_version 2350 (0.0007)
[2023-03-03 11:33:17,685][16994] Fps is (10 sec: 3174.2, 60 sec: 3328.0, 300 sec: 3061.6). Total num frames: 2412544. Throughput: 0: 3357.8. Samples: 2412128. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:33:17,686][16994] Avg episode reward: [(0, '23.482')]
[2023-03-03 11:33:18,826][17031] Updated weights for policy 0, policy_version 2360 (0.0007)
[2023-03-03 11:33:22,445][17031] Updated weights for policy 0, policy_version 2370 (0.0009)
[2023-03-03 11:33:22,684][16994] Fps is (10 sec: 2968.6, 60 sec: 3293.8, 300 sec: 3065.1). Total num frames: 2426880. Throughput: 0: 3324.3. Samples: 2420382. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:33:22,685][16994] Avg episode reward: [(0, '26.256')]
[2023-03-03 11:33:25,630][17031] Updated weights for policy 0, policy_version 2380 (0.0006)
[2023-03-03 11:33:27,684][16994] Fps is (10 sec: 3072.1, 60 sec: 3293.7, 300 sec: 3072.0). Total num frames: 2443264. Throughput: 0: 3283.0. Samples: 2438899. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:33:27,685][16994] Avg episode reward: [(0, '26.117')]
[2023-03-03 11:33:28,897][17031] Updated weights for policy 0, policy_version 2390 (0.0007)
[2023-03-03 11:33:32,437][17031] Updated weights for policy 0, policy_version 2400 (0.0007)
[2023-03-03 11:33:32,683][16994] Fps is (10 sec: 3072.4, 60 sec: 3259.8, 300 sec: 3072.0). Total num frames: 2457600. Throughput: 0: 3226.0. Samples: 2456836. Policy #0 lag: (min: 0.0, avg: 0.9, max: 1.0)
[2023-03-03 11:33:32,684][16994] Avg episode reward: [(0, '31.739')]
[2023-03-03 11:33:35,962][17031] Updated weights for policy 0, policy_version 2410 (0.0007)
[2023-03-03 11:33:37,685][16994] Fps is (10 sec: 2866.9, 60 sec: 3225.6, 300 sec: 3068.5). Total num frames: 2471936. Throughput: 0: 3190.1. Samples: 2465467. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:33:37,686][16994] Avg episode reward: [(0, '29.308')]
[2023-03-03 11:33:37,755][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002415_2472960.pth...
[2023-03-03 11:33:37,830][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000001691_1731584.pth
[2023-03-03 11:33:39,378][17031] Updated weights for policy 0, policy_version 2420 (0.0008)
[2023-03-03 11:33:42,594][17031] Updated weights for policy 0, policy_version 2430 (0.0007)
[2023-03-03 11:33:42,680][16994] Fps is (10 sec: 3072.9, 60 sec: 3208.7, 300 sec: 3068.6). Total num frames: 2488320. Throughput: 0: 3162.3. Samples: 2484180. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:33:42,680][16994] Avg episode reward: [(0, '33.102')]
[2023-03-03 11:33:45,841][17031] Updated weights for policy 0, policy_version 2440 (0.0008)
[2023-03-03 11:33:47,683][16994] Fps is (10 sec: 3175.1, 60 sec: 3174.2, 300 sec: 3072.0). Total num frames: 2503680. Throughput: 0: 3130.8. Samples: 2503042. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:33:47,683][16994] Avg episode reward: [(0, '31.296')]
[2023-03-03 11:33:48,869][17031] Updated weights for policy 0, policy_version 2450 (0.0006)
[2023-03-03 11:33:52,082][17031] Updated weights for policy 0, policy_version 2460 (0.0007)
[2023-03-03 11:33:52,684][16994] Fps is (10 sec: 3173.3, 60 sec: 3174.4, 300 sec: 3078.9). Total num frames: 2520064. Throughput: 0: 3118.0. Samples: 2512658. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:33:52,685][16994] Avg episode reward: [(0, '32.047')]
[2023-03-03 11:33:55,386][17031] Updated weights for policy 0, policy_version 2470 (0.0006)
[2023-03-03 11:33:57,683][16994] Fps is (10 sec: 3276.9, 60 sec: 3157.3, 300 sec: 3089.4). Total num frames: 2536448. Throughput: 0: 3093.9. Samples: 2531914. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:33:57,683][16994] Avg episode reward: [(0, '27.344')]
[2023-03-03 11:33:58,432][17031] Updated weights for policy 0, policy_version 2480 (0.0006)
[2023-03-03 11:34:01,465][17031] Updated weights for policy 0, policy_version 2490 (0.0007)
[2023-03-03 11:34:02,684][16994] Fps is (10 sec: 3379.0, 60 sec: 3174.4, 300 sec: 3096.3). Total num frames: 2553856. Throughput: 0: 3110.3. Samples: 2552091. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:02,684][16994] Avg episode reward: [(0, '24.376')]
[2023-03-03 11:34:04,531][17031] Updated weights for policy 0, policy_version 2500 (0.0007)
[2023-03-03 11:34:07,684][16994] Fps is (10 sec: 3276.4, 60 sec: 3140.2, 300 sec: 3096.3). Total num frames: 2569216. Throughput: 0: 3147.4. Samples: 2562016. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:34:07,685][16994] Avg episode reward: [(0, '22.074')]
[2023-03-03 11:34:07,710][17031] Updated weights for policy 0, policy_version 2510 (0.0008)
[2023-03-03 11:34:11,112][17031] Updated weights for policy 0, policy_version 2520 (0.0009)
[2023-03-03 11:34:12,685][16994] Fps is (10 sec: 3071.7, 60 sec: 3123.0, 300 sec: 3099.7). Total num frames: 2584576. Throughput: 0: 3148.6. Samples: 2580589. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:12,687][16994] Avg episode reward: [(0, '24.059')]
[2023-03-03 11:34:14,603][17031] Updated weights for policy 0, policy_version 2530 (0.0006)
[2023-03-03 11:34:17,683][16994] Fps is (10 sec: 3072.1, 60 sec: 3123.2, 300 sec: 3106.7). Total num frames: 2599936. Throughput: 0: 3151.8. Samples: 2598670. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:17,684][16994] Avg episode reward: [(0, '23.265')]
[2023-03-03 11:34:17,778][17031] Updated weights for policy 0, policy_version 2540 (0.0008)
[2023-03-03 11:34:20,789][17031] Updated weights for policy 0, policy_version 2550 (0.0006)
[2023-03-03 11:34:22,682][16994] Fps is (10 sec: 3175.4, 60 sec: 3157.4, 300 sec: 3110.2). Total num frames: 2616320. Throughput: 0: 3186.3. Samples: 2608842. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:22,683][16994] Avg episode reward: [(0, '24.341')]
[2023-03-03 11:34:23,855][17031] Updated weights for policy 0, policy_version 2560 (0.0007)
[2023-03-03 11:34:27,090][17031] Updated weights for policy 0, policy_version 2570 (0.0006)
[2023-03-03 11:34:27,684][16994] Fps is (10 sec: 3276.7, 60 sec: 3157.3, 300 sec: 3110.2). Total num frames: 2632704. Throughput: 0: 3206.5. Samples: 2628484. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:34:27,684][16994] Avg episode reward: [(0, '22.922')]
[2023-03-03 11:34:30,220][17031] Updated weights for policy 0, policy_version 2580 (0.0008)
[2023-03-03 11:34:32,683][16994] Fps is (10 sec: 3378.8, 60 sec: 3208.5, 300 sec: 3117.1). Total num frames: 2650112. Throughput: 0: 3232.5. Samples: 2648503. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:34:32,684][16994] Avg episode reward: [(0, '23.665')]
[2023-03-03 11:34:33,271][17031] Updated weights for policy 0, policy_version 2590 (0.0007)
[2023-03-03 11:34:36,515][17031] Updated weights for policy 0, policy_version 2600 (0.0007)
[2023-03-03 11:34:37,681][16994] Fps is (10 sec: 3277.7, 60 sec: 3225.8, 300 sec: 3113.7). Total num frames: 2665472. Throughput: 0: 3244.7. Samples: 2658661. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:34:37,682][16994] Avg episode reward: [(0, '21.654')]
[2023-03-03 11:34:39,973][17031] Updated weights for policy 0, policy_version 2610 (0.0006)
[2023-03-03 11:34:42,681][16994] Fps is (10 sec: 2970.2, 60 sec: 3191.4, 300 sec: 3110.2). Total num frames: 2679808. Throughput: 0: 3181.5. Samples: 2675076. Policy #0 lag: (min: 0.0, avg: 0.7, max: 2.0)
[2023-03-03 11:34:42,683][16994] Avg episode reward: [(0, '22.080')]
[2023-03-03 11:34:43,588][17031] Updated weights for policy 0, policy_version 2620 (0.0009)
[2023-03-03 11:34:46,725][17031] Updated weights for policy 0, policy_version 2630 (0.0007)
[2023-03-03 11:34:47,683][16994] Fps is (10 sec: 3071.3, 60 sec: 3208.5, 300 sec: 3113.7). Total num frames: 2696192. Throughput: 0: 3166.7. Samples: 2694590. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:47,684][16994] Avg episode reward: [(0, '23.732')]
[2023-03-03 11:34:49,751][17031] Updated weights for policy 0, policy_version 2640 (0.0007)
[2023-03-03 11:34:52,683][16994] Fps is (10 sec: 3276.0, 60 sec: 3208.5, 300 sec: 3113.7). Total num frames: 2712576. Throughput: 0: 3169.7. Samples: 2704653. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:52,684][16994] Avg episode reward: [(0, '26.436')]
[2023-03-03 11:34:52,822][17031] Updated weights for policy 0, policy_version 2650 (0.0006)
[2023-03-03 11:34:55,859][17031] Updated weights for policy 0, policy_version 2660 (0.0006)
[2023-03-03 11:34:57,681][16994] Fps is (10 sec: 3277.6, 60 sec: 3208.6, 300 sec: 3113.7). Total num frames: 2728960. Throughput: 0: 3204.6. Samples: 2724783. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:34:57,681][16994] Avg episode reward: [(0, '26.730')]
[2023-03-03 11:34:58,889][17031] Updated weights for policy 0, policy_version 2670 (0.0007)
[2023-03-03 11:35:01,922][17031] Updated weights for policy 0, policy_version 2680 (0.0006)
[2023-03-03 11:35:02,683][16994] Fps is (10 sec: 3379.3, 60 sec: 3208.6, 300 sec: 3127.5). Total num frames: 2746368. Throughput: 0: 3253.1. Samples: 2745058. Policy #0 lag: (min: 0.0, avg: 0.9, max: 2.0)
[2023-03-03 11:35:02,684][16994] Avg episode reward: [(0, '25.389')]
[2023-03-03 11:35:04,973][17031] Updated weights for policy 0, policy_version 2690 (0.0007)
[2023-03-03 11:35:07,683][16994] Fps is (10 sec: 3276.0, 60 sec: 3208.6, 300 sec: 3127.6). Total num frames: 2761728. Throughput: 0: 3250.6. Samples: 2755125. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:35:07,684][16994] Avg episode reward: [(0, '23.020')]
[2023-03-03 11:35:08,645][17031] Updated weights for policy 0, policy_version 2700 (0.0007)
[2023-03-03 11:35:11,945][17031] Updated weights for policy 0, policy_version 2710 (0.0007)
[2023-03-03 11:35:12,681][16994] Fps is (10 sec: 3072.7, 60 sec: 3208.8, 300 sec: 3131.1). Total num frames: 2777088. Throughput: 0: 3201.6. Samples: 2772544. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:35:12,681][16994] Avg episode reward: [(0, '23.368')]
[2023-03-03 11:35:15,370][17031] Updated weights for policy 0, policy_version 2720 (0.0008)
[2023-03-03 11:35:17,682][16994] Fps is (10 sec: 2969.9, 60 sec: 3191.5, 300 sec: 3131.0). Total num frames: 2791424. Throughput: 0: 3157.8. Samples: 2790602. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:35:17,684][16994] Avg episode reward: [(0, '26.126')]
[2023-03-03 11:35:18,877][17031] Updated weights for policy 0, policy_version 2730 (0.0008)
[2023-03-03 11:35:22,233][17031] Updated weights for policy 0, policy_version 2740 (0.0007)
[2023-03-03 11:35:22,680][16994] Fps is (10 sec: 2969.7, 60 sec: 3174.5, 300 sec: 3134.5). Total num frames: 2806784. Throughput: 0: 3132.4. Samples: 2799615. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:35:22,681][16994] Avg episode reward: [(0, '31.719')]
[2023-03-03 11:35:25,359][17031] Updated weights for policy 0, policy_version 2750 (0.0006)
[2023-03-03 11:35:27,680][16994] Fps is (10 sec: 3175.1, 60 sec: 3174.6, 300 sec: 3141.4). Total num frames: 2823168. Throughput: 0: 3189.9. Samples: 2818617. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:35:27,681][16994] Avg episode reward: [(0, '35.354')]
[2023-03-03 11:35:28,514][17031] Updated weights for policy 0, policy_version 2760 (0.0007)
[2023-03-03 11:35:31,757][17031] Updated weights for policy 0, policy_version 2770 (0.0006)
[2023-03-03 11:35:32,681][16994] Fps is (10 sec: 3174.2, 60 sec: 3140.4, 300 sec: 3145.0). Total num frames: 2838528. Throughput: 0: 3180.6. Samples: 2837712. Policy #0 lag: (min: 0.0, avg: 0.8, max: 1.0)
[2023-03-03 11:35:32,683][16994] Avg episode reward: [(0, '33.391')]
[2023-03-03 11:35:35,204][17031] Updated weights for policy 0, policy_version 2780 (0.0008)
[2023-03-03 11:35:37,684][16994] Fps is (10 sec: 2968.5, 60 sec: 3123.1, 300 sec: 3144.9). Total num frames: 2852864. Throughput: 0: 3151.4. Samples: 2846466. Policy #0 lag: (min: 0.0, avg: 0.8, max: 2.0)
[2023-03-03 11:35:37,688][16994] Avg episode reward: [(0, '28.531')]
[2023-03-03 11:35:37,694][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002787_2853888.pth...
[2023-03-03 11:35:37,788][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002044_2093056.pth
[2023-03-03 11:35:38,865][17031] Updated weights for policy 0, policy_version 2790 (0.0009)
[2023-03-03 11:35:42,421][17031] Updated weights for policy 0, policy_version 2800 (0.0008)
[2023-03-03 11:35:42,685][16994] Fps is (10 sec: 2866.4, 60 sec: 3123.1, 300 sec: 3141.4). Total num frames: 2867200. Throughput: 0: 3074.9. Samples: 2863164. Policy #0 lag: (min: 0.0, avg: 0.7, max: 1.0)
[2023-03-03 11:35:42,692][16994] Avg episode reward: [(0, '26.635')]
[2023-03-03 11:35:44,767][16994] Keyboard interrupt detected in the event loop EvtLoop [Runner_EvtLoop, process=main process 16994], exiting...
[2023-03-03 11:35:44,776][16994] Runner profile tree view:
main_loop: 965.0570
[2023-03-03 11:35:44,779][16994] Collected {0: 2873344}, FPS: 2977.4
[2023-03-03 11:35:44,770][17030] Stopping Batcher_0...
[2023-03-03 11:35:44,781][17030] Loop batcher_evt_loop terminating...
[2023-03-03 11:35:44,784][17030] Saving /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002806_2873344.pth...
[2023-03-03 11:35:44,774][17036] Stopping RolloutWorker_w5...
[2023-03-03 11:35:44,776][17035] Stopping RolloutWorker_w3...
[2023-03-03 11:35:44,787][17036] Loop rollout_proc5_evt_loop terminating...
[2023-03-03 11:35:44,789][17035] Loop rollout_proc3_evt_loop terminating...
[2023-03-03 11:35:44,774][17033] Stopping RolloutWorker_w1...
[2023-03-03 11:35:44,777][17034] Stopping RolloutWorker_w2...
[2023-03-03 11:35:44,791][17033] Loop rollout_proc1_evt_loop terminating...
[2023-03-03 11:35:44,792][17034] Loop rollout_proc2_evt_loop terminating...
[2023-03-03 11:35:44,774][17037] Stopping RolloutWorker_w4...
[2023-03-03 11:35:44,780][17038] Stopping RolloutWorker_w6...
[2023-03-03 11:35:44,794][17038] Loop rollout_proc6_evt_loop terminating...
[2023-03-03 11:35:44,794][17037] Loop rollout_proc4_evt_loop terminating...
[2023-03-03 11:35:44,790][17032] Stopping RolloutWorker_w0...
[2023-03-03 11:35:44,789][17039] Stopping RolloutWorker_w7...
[2023-03-03 11:35:44,802][17032] Loop rollout_proc0_evt_loop terminating...
[2023-03-03 11:35:44,805][17039] Loop rollout_proc7_evt_loop terminating...
[2023-03-03 11:35:44,917][17030] Removing /Users/quentingallouedec/gia/data/envs/metaworld/train_dir/pick-place-v2/checkpoint_p0/checkpoint_000002415_2472960.pth
[2023-03-03 11:35:44,958][17030] Stopping LearnerWorker_p0...
[2023-03-03 11:35:44,958][17030] Loop learner_proc0_evt_loop terminating...
[2023-03-03 11:35:45,199][17031] Weights refcount: 2 0
[2023-03-03 11:35:45,205][17031] Stopping InferenceWorker_p0-w0...
[2023-03-03 11:35:45,206][17031] Loop inference_proc0-0_evt_loop terminating...