Taratata commited on
Commit
ab3246b
1 Parent(s): e738185

Upload . with huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. replay.mp4 +2 -2
  3. sf_log.txt +282 -0
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 8.10 +/- 3.73
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 9.48 +/- 4.69
19
  name: mean_reward
20
  verified: false
21
  ---
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85544d48b86c629f4a0a237f1d95ca2e788ff0fd8f91aaaca5e4ce0b98a4a76c
3
- size 14721558
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b1808132fd6f5e6864ab3695bf2ec741b1bb2be819e19c7e32c48a86b1d0593
3
+ size 17548925
sf_log.txt CHANGED
@@ -1283,3 +1283,285 @@ main_loop: 1180.8338
1283
  [2023-03-11 11:45:53,574][00127] Avg episode rewards: #0: 17.000, true rewards: #0: 8.100
1284
  [2023-03-11 11:45:53,575][00127] Avg episode reward: 17.000, avg true_objective: 8.100
1285
  [2023-03-11 11:46:41,540][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1283
  [2023-03-11 11:45:53,574][00127] Avg episode rewards: #0: 17.000, true rewards: #0: 8.100
1284
  [2023-03-11 11:45:53,575][00127] Avg episode reward: 17.000, avg true_objective: 8.100
1285
  [2023-03-11 11:46:41,540][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
1286
+ [2023-03-11 11:46:53,587][00127] The model has been pushed to https://huggingface.co/Taratata/rl_course_vizdoom_health_gathering_supreme
1287
+ [2023-03-11 11:48:24,968][00127] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
1288
+ [2023-03-11 11:48:24,970][00127] Overriding arg 'num_workers' with value 1 passed from command line
1289
+ [2023-03-11 11:48:24,972][00127] Adding new argument 'no_render'=True that is not in the saved config file!
1290
+ [2023-03-11 11:48:24,974][00127] Adding new argument 'save_video'=True that is not in the saved config file!
1291
+ [2023-03-11 11:48:24,976][00127] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1292
+ [2023-03-11 11:48:24,977][00127] Adding new argument 'video_name'=None that is not in the saved config file!
1293
+ [2023-03-11 11:48:24,979][00127] Adding new argument 'max_num_frames'=1000000000.0 that is not in the saved config file!
1294
+ [2023-03-11 11:48:24,981][00127] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1295
+ [2023-03-11 11:48:24,982][00127] Adding new argument 'push_to_hub'=False that is not in the saved config file!
1296
+ [2023-03-11 11:48:24,983][00127] Adding new argument 'hf_repository'=None that is not in the saved config file!
1297
+ [2023-03-11 11:48:24,984][00127] Adding new argument 'policy_index'=0 that is not in the saved config file!
1298
+ [2023-03-11 11:48:24,985][00127] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1299
+ [2023-03-11 11:48:24,986][00127] Adding new argument 'train_script'=None that is not in the saved config file!
1300
+ [2023-03-11 11:48:24,987][00127] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1301
+ [2023-03-11 11:48:24,988][00127] Using frameskip 1 and render_action_repeat=4 for evaluation
1302
+ [2023-03-11 11:48:25,005][00127] RunningMeanStd input shape: (3, 72, 128)
1303
+ [2023-03-11 11:48:25,008][00127] RunningMeanStd input shape: (1,)
1304
+ [2023-03-11 11:48:25,022][00127] ConvEncoder: input_channels=3
1305
+ [2023-03-11 11:48:25,062][00127] Conv encoder output size: 512
1306
+ [2023-03-11 11:48:25,063][00127] Policy head output size: 512
1307
+ [2023-03-11 11:48:25,082][00127] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000978_4005888.pth...
1308
+ [2023-03-11 11:48:25,531][00127] Num frames 100...
1309
+ [2023-03-11 11:48:25,643][00127] Num frames 200...
1310
+ [2023-03-11 11:48:25,755][00127] Num frames 300...
1311
+ [2023-03-11 11:48:25,864][00127] Num frames 400...
1312
+ [2023-03-11 11:48:25,984][00127] Num frames 500...
1313
+ [2023-03-11 11:48:26,098][00127] Num frames 600...
1314
+ [2023-03-11 11:48:26,204][00127] Num frames 700...
1315
+ [2023-03-11 11:48:26,316][00127] Num frames 800...
1316
+ [2023-03-11 11:48:26,435][00127] Num frames 900...
1317
+ [2023-03-11 11:48:26,552][00127] Num frames 1000...
1318
+ [2023-03-11 11:48:26,668][00127] Avg episode rewards: #0: 22.550, true rewards: #0: 10.550
1319
+ [2023-03-11 11:48:26,671][00127] Avg episode reward: 22.550, avg true_objective: 10.550
1320
+ [2023-03-11 11:48:26,724][00127] Num frames 1100...
1321
+ [2023-03-11 11:48:26,841][00127] Num frames 1200...
1322
+ [2023-03-11 11:48:26,952][00127] Num frames 1300...
1323
+ [2023-03-11 11:48:27,062][00127] Num frames 1400...
1324
+ [2023-03-11 11:48:27,171][00127] Num frames 1500...
1325
+ [2023-03-11 11:48:27,288][00127] Num frames 1600...
1326
+ [2023-03-11 11:48:27,410][00127] Num frames 1700...
1327
+ [2023-03-11 11:48:27,525][00127] Num frames 1800...
1328
+ [2023-03-11 11:48:27,645][00127] Num frames 1900...
1329
+ [2023-03-11 11:48:27,754][00127] Num frames 2000...
1330
+ [2023-03-11 11:48:27,864][00127] Num frames 2100...
1331
+ [2023-03-11 11:48:27,986][00127] Num frames 2200...
1332
+ [2023-03-11 11:48:28,109][00127] Num frames 2300...
1333
+ [2023-03-11 11:48:28,220][00127] Avg episode rewards: #0: 27.245, true rewards: #0: 11.745
1334
+ [2023-03-11 11:48:28,222][00127] Avg episode reward: 27.245, avg true_objective: 11.745
1335
+ [2023-03-11 11:48:28,284][00127] Num frames 2400...
1336
+ [2023-03-11 11:48:28,395][00127] Num frames 2500...
1337
+ [2023-03-11 11:48:28,507][00127] Num frames 2600...
1338
+ [2023-03-11 11:48:28,624][00127] Num frames 2700...
1339
+ [2023-03-11 11:48:28,747][00127] Num frames 2800...
1340
+ [2023-03-11 11:48:28,858][00127] Num frames 2900...
1341
+ [2023-03-11 11:48:28,968][00127] Num frames 3000...
1342
+ [2023-03-11 11:48:29,074][00127] Num frames 3100...
1343
+ [2023-03-11 11:48:29,186][00127] Num frames 3200...
1344
+ [2023-03-11 11:48:29,301][00127] Num frames 3300...
1345
+ [2023-03-11 11:48:29,425][00127] Num frames 3400...
1346
+ [2023-03-11 11:48:29,535][00127] Num frames 3500...
1347
+ [2023-03-11 11:48:29,652][00127] Num frames 3600...
1348
+ [2023-03-11 11:48:29,764][00127] Num frames 3700...
1349
+ [2023-03-11 11:48:29,879][00127] Num frames 3800...
1350
+ [2023-03-11 11:48:29,989][00127] Num frames 3900...
1351
+ [2023-03-11 11:48:30,118][00127] Num frames 4000...
1352
+ [2023-03-11 11:48:30,231][00127] Num frames 4100...
1353
+ [2023-03-11 11:48:30,340][00127] Num frames 4200...
1354
+ [2023-03-11 11:48:30,472][00127] Avg episode rewards: #0: 35.563, true rewards: #0: 14.230
1355
+ [2023-03-11 11:48:30,474][00127] Avg episode reward: 35.563, avg true_objective: 14.230
1356
+ [2023-03-11 11:48:30,517][00127] Num frames 4300...
1357
+ [2023-03-11 11:48:30,634][00127] Num frames 4400...
1358
+ [2023-03-11 11:48:30,753][00127] Num frames 4500...
1359
+ [2023-03-11 11:48:30,864][00127] Num frames 4600...
1360
+ [2023-03-11 11:48:30,978][00127] Num frames 4700...
1361
+ [2023-03-11 11:48:31,088][00127] Num frames 4800...
1362
+ [2023-03-11 11:48:31,199][00127] Num frames 4900...
1363
+ [2023-03-11 11:48:31,310][00127] Num frames 5000...
1364
+ [2023-03-11 11:48:31,408][00127] Avg episode rewards: #0: 30.592, true rewards: #0: 12.592
1365
+ [2023-03-11 11:48:31,410][00127] Avg episode reward: 30.592, avg true_objective: 12.592
1366
+ [2023-03-11 11:48:31,481][00127] Num frames 5100...
1367
+ [2023-03-11 11:48:31,590][00127] Num frames 5200...
1368
+ [2023-03-11 11:48:31,710][00127] Num frames 5300...
1369
+ [2023-03-11 11:48:31,822][00127] Num frames 5400...
1370
+ [2023-03-11 11:48:31,931][00127] Num frames 5500...
1371
+ [2023-03-11 11:48:32,062][00127] Num frames 5600...
1372
+ [2023-03-11 11:48:32,169][00127] Num frames 5700...
1373
+ [2023-03-11 11:48:32,284][00127] Num frames 5800...
1374
+ [2023-03-11 11:48:32,398][00127] Num frames 5900...
1375
+ [2023-03-11 11:48:32,513][00127] Num frames 6000...
1376
+ [2023-03-11 11:48:32,621][00127] Num frames 6100...
1377
+ [2023-03-11 11:48:32,743][00127] Num frames 6200...
1378
+ [2023-03-11 11:48:32,862][00127] Num frames 6300...
1379
+ [2023-03-11 11:48:32,959][00127] Avg episode rewards: #0: 30.476, true rewards: #0: 12.676
1380
+ [2023-03-11 11:48:32,960][00127] Avg episode reward: 30.476, avg true_objective: 12.676
1381
+ [2023-03-11 11:48:33,029][00127] Num frames 6400...
1382
+ [2023-03-11 11:48:33,138][00127] Num frames 6500...
1383
+ [2023-03-11 11:48:33,295][00127] Num frames 6600...
1384
+ [2023-03-11 11:48:33,485][00127] Avg episode rewards: #0: 26.150, true rewards: #0: 11.150
1385
+ [2023-03-11 11:48:33,487][00127] Avg episode reward: 26.150, avg true_objective: 11.150
1386
+ [2023-03-11 11:48:33,516][00127] Num frames 6700...
1387
+ [2023-03-11 11:48:33,686][00127] Num frames 6800...
1388
+ [2023-03-11 11:48:33,842][00127] Num frames 6900...
1389
+ [2023-03-11 11:48:33,997][00127] Num frames 7000...
1390
+ [2023-03-11 11:48:34,151][00127] Num frames 7100...
1391
+ [2023-03-11 11:48:34,303][00127] Num frames 7200...
1392
+ [2023-03-11 11:48:34,460][00127] Num frames 7300...
1393
+ [2023-03-11 11:48:34,612][00127] Num frames 7400...
1394
+ [2023-03-11 11:48:34,786][00127] Num frames 7500...
1395
+ [2023-03-11 11:48:34,936][00127] Num frames 7600...
1396
+ [2023-03-11 11:48:35,092][00127] Num frames 7700...
1397
+ [2023-03-11 11:48:35,266][00127] Avg episode rewards: #0: 26.254, true rewards: #0: 11.111
1398
+ [2023-03-11 11:48:35,268][00127] Avg episode reward: 26.254, avg true_objective: 11.111
1399
+ [2023-03-11 11:48:35,311][00127] Num frames 7800...
1400
+ [2023-03-11 11:48:35,470][00127] Num frames 7900...
1401
+ [2023-03-11 11:48:35,629][00127] Num frames 8000...
1402
+ [2023-03-11 11:48:35,788][00127] Num frames 8100...
1403
+ [2023-03-11 11:48:35,899][00127] Avg episode rewards: #0: 23.542, true rewards: #0: 10.167
1404
+ [2023-03-11 11:48:35,901][00127] Avg episode reward: 23.542, avg true_objective: 10.167
1405
+ [2023-03-11 11:48:36,004][00127] Num frames 8200...
1406
+ [2023-03-11 11:48:36,157][00127] Num frames 8300...
1407
+ [2023-03-11 11:48:36,312][00127] Num frames 8400...
1408
+ [2023-03-11 11:48:36,466][00127] Num frames 8500...
1409
+ [2023-03-11 11:48:36,632][00127] Num frames 8600...
1410
+ [2023-03-11 11:48:36,761][00127] Num frames 8700...
1411
+ [2023-03-11 11:48:36,879][00127] Num frames 8800...
1412
+ [2023-03-11 11:48:36,997][00127] Num frames 8900...
1413
+ [2023-03-11 11:48:37,115][00127] Num frames 9000...
1414
+ [2023-03-11 11:48:37,223][00127] Num frames 9100...
1415
+ [2023-03-11 11:48:37,335][00127] Num frames 9200...
1416
+ [2023-03-11 11:48:37,492][00127] Avg episode rewards: #0: 23.762, true rewards: #0: 10.318
1417
+ [2023-03-11 11:48:37,494][00127] Avg episode reward: 23.762, avg true_objective: 10.318
1418
+ [2023-03-11 11:48:37,514][00127] Num frames 9300...
1419
+ [2023-03-11 11:48:37,624][00127] Num frames 9400...
1420
+ [2023-03-11 11:48:37,732][00127] Num frames 9500...
1421
+ [2023-03-11 11:48:37,849][00127] Num frames 9600...
1422
+ [2023-03-11 11:48:37,962][00127] Num frames 9700...
1423
+ [2023-03-11 11:48:38,073][00127] Num frames 9800...
1424
+ [2023-03-11 11:48:38,183][00127] Num frames 9900...
1425
+ [2023-03-11 11:48:38,301][00127] Num frames 10000...
1426
+ [2023-03-11 11:48:38,411][00127] Num frames 10100...
1427
+ [2023-03-11 11:48:38,520][00127] Num frames 10200...
1428
+ [2023-03-11 11:48:38,627][00127] Num frames 10300...
1429
+ [2023-03-11 11:48:38,732][00127] Avg episode rewards: #0: 23.841, true rewards: #0: 10.341
1430
+ [2023-03-11 11:48:38,733][00127] Avg episode reward: 23.841, avg true_objective: 10.341
1431
+ [2023-03-11 11:49:39,627][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
1432
+ [2023-03-11 11:50:02,077][00127] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
1433
+ [2023-03-11 11:50:02,079][00127] Overriding arg 'num_workers' with value 1 passed from command line
1434
+ [2023-03-11 11:50:02,081][00127] Adding new argument 'no_render'=True that is not in the saved config file!
1435
+ [2023-03-11 11:50:02,083][00127] Adding new argument 'save_video'=True that is not in the saved config file!
1436
+ [2023-03-11 11:50:02,085][00127] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1437
+ [2023-03-11 11:50:02,087][00127] Adding new argument 'video_name'=None that is not in the saved config file!
1438
+ [2023-03-11 11:50:02,088][00127] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
1439
+ [2023-03-11 11:50:02,090][00127] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1440
+ [2023-03-11 11:50:02,091][00127] Adding new argument 'push_to_hub'=True that is not in the saved config file!
1441
+ [2023-03-11 11:50:02,092][00127] Adding new argument 'hf_repository'='Taratata/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
1442
+ [2023-03-11 11:50:02,093][00127] Adding new argument 'policy_index'=0 that is not in the saved config file!
1443
+ [2023-03-11 11:50:02,094][00127] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1444
+ [2023-03-11 11:50:02,095][00127] Adding new argument 'train_script'=None that is not in the saved config file!
1445
+ [2023-03-11 11:50:02,096][00127] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1446
+ [2023-03-11 11:50:02,097][00127] Using frameskip 1 and render_action_repeat=4 for evaluation
1447
+ [2023-03-11 11:50:02,117][00127] RunningMeanStd input shape: (3, 72, 128)
1448
+ [2023-03-11 11:50:02,120][00127] RunningMeanStd input shape: (1,)
1449
+ [2023-03-11 11:50:02,133][00127] ConvEncoder: input_channels=3
1450
+ [2023-03-11 11:50:02,168][00127] Conv encoder output size: 512
1451
+ [2023-03-11 11:50:02,172][00127] Policy head output size: 512
1452
+ [2023-03-11 11:50:02,192][00127] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000978_4005888.pth...
1453
+ [2023-03-11 11:50:02,628][00127] Num frames 100...
1454
+ [2023-03-11 11:50:02,748][00127] Num frames 200...
1455
+ [2023-03-11 11:50:02,868][00127] Num frames 300...
1456
+ [2023-03-11 11:50:02,990][00127] Num frames 400...
1457
+ [2023-03-11 11:50:03,108][00127] Num frames 500...
1458
+ [2023-03-11 11:50:03,227][00127] Num frames 600...
1459
+ [2023-03-11 11:50:03,340][00127] Num frames 700...
1460
+ [2023-03-11 11:50:03,459][00127] Num frames 800...
1461
+ [2023-03-11 11:50:03,580][00127] Num frames 900...
1462
+ [2023-03-11 11:50:03,696][00127] Num frames 1000...
1463
+ [2023-03-11 11:50:03,768][00127] Avg episode rewards: #0: 22.080, true rewards: #0: 10.080
1464
+ [2023-03-11 11:50:03,770][00127] Avg episode reward: 22.080, avg true_objective: 10.080
1465
+ [2023-03-11 11:50:03,875][00127] Num frames 1100...
1466
+ [2023-03-11 11:50:03,999][00127] Num frames 1200...
1467
+ [2023-03-11 11:50:04,108][00127] Num frames 1300...
1468
+ [2023-03-11 11:50:04,224][00127] Num frames 1400...
1469
+ [2023-03-11 11:50:04,339][00127] Num frames 1500...
1470
+ [2023-03-11 11:50:04,450][00127] Num frames 1600...
1471
+ [2023-03-11 11:50:04,566][00127] Num frames 1700...
1472
+ [2023-03-11 11:50:04,685][00127] Num frames 1800...
1473
+ [2023-03-11 11:50:04,801][00127] Num frames 1900...
1474
+ [2023-03-11 11:50:04,909][00127] Num frames 2000...
1475
+ [2023-03-11 11:50:05,022][00127] Num frames 2100...
1476
+ [2023-03-11 11:50:05,132][00127] Num frames 2200...
1477
+ [2023-03-11 11:50:05,247][00127] Num frames 2300...
1478
+ [2023-03-11 11:50:05,357][00127] Num frames 2400...
1479
+ [2023-03-11 11:50:05,469][00127] Num frames 2500...
1480
+ [2023-03-11 11:50:05,580][00127] Num frames 2600...
1481
+ [2023-03-11 11:50:05,695][00127] Num frames 2700...
1482
+ [2023-03-11 11:50:05,807][00127] Num frames 2800...
1483
+ [2023-03-11 11:50:05,859][00127] Avg episode rewards: #0: 32.500, true rewards: #0: 14.000
1484
+ [2023-03-11 11:50:05,861][00127] Avg episode reward: 32.500, avg true_objective: 14.000
1485
+ [2023-03-11 11:50:05,977][00127] Num frames 2900...
1486
+ [2023-03-11 11:50:06,085][00127] Num frames 3000...
1487
+ [2023-03-11 11:50:06,194][00127] Num frames 3100...
1488
+ [2023-03-11 11:50:06,311][00127] Num frames 3200...
1489
+ [2023-03-11 11:50:06,423][00127] Num frames 3300...
1490
+ [2023-03-11 11:50:06,531][00127] Num frames 3400...
1491
+ [2023-03-11 11:50:06,661][00127] Num frames 3500...
1492
+ [2023-03-11 11:50:06,776][00127] Num frames 3600...
1493
+ [2023-03-11 11:50:06,828][00127] Avg episode rewards: #0: 27.333, true rewards: #0: 12.000
1494
+ [2023-03-11 11:50:06,830][00127] Avg episode reward: 27.333, avg true_objective: 12.000
1495
+ [2023-03-11 11:50:06,941][00127] Num frames 3700...
1496
+ [2023-03-11 11:50:07,058][00127] Num frames 3800...
1497
+ [2023-03-11 11:50:07,163][00127] Num frames 3900...
1498
+ [2023-03-11 11:50:07,270][00127] Num frames 4000...
1499
+ [2023-03-11 11:50:07,381][00127] Avg episode rewards: #0: 21.870, true rewards: #0: 10.120
1500
+ [2023-03-11 11:50:07,383][00127] Avg episode reward: 21.870, avg true_objective: 10.120
1501
+ [2023-03-11 11:50:07,452][00127] Num frames 4100...
1502
+ [2023-03-11 11:50:07,565][00127] Num frames 4200...
1503
+ [2023-03-11 11:50:07,679][00127] Num frames 4300...
1504
+ [2023-03-11 11:50:07,790][00127] Num frames 4400...
1505
+ [2023-03-11 11:50:07,908][00127] Num frames 4500...
1506
+ [2023-03-11 11:50:08,029][00127] Num frames 4600...
1507
+ [2023-03-11 11:50:08,141][00127] Num frames 4700...
1508
+ [2023-03-11 11:50:08,257][00127] Num frames 4800...
1509
+ [2023-03-11 11:50:08,378][00127] Num frames 4900...
1510
+ [2023-03-11 11:50:08,483][00127] Avg episode rewards: #0: 21.488, true rewards: #0: 9.888
1511
+ [2023-03-11 11:50:08,484][00127] Avg episode reward: 21.488, avg true_objective: 9.888
1512
+ [2023-03-11 11:50:08,553][00127] Num frames 5000...
1513
+ [2023-03-11 11:50:08,665][00127] Num frames 5100...
1514
+ [2023-03-11 11:50:08,776][00127] Num frames 5200...
1515
+ [2023-03-11 11:50:08,905][00127] Num frames 5300...
1516
+ [2023-03-11 11:50:09,033][00127] Num frames 5400...
1517
+ [2023-03-11 11:50:09,144][00127] Num frames 5500...
1518
+ [2023-03-11 11:50:09,255][00127] Num frames 5600...
1519
+ [2023-03-11 11:50:09,398][00127] Avg episode rewards: #0: 20.462, true rewards: #0: 9.462
1520
+ [2023-03-11 11:50:09,400][00127] Avg episode reward: 20.462, avg true_objective: 9.462
1521
+ [2023-03-11 11:50:09,440][00127] Num frames 5700...
1522
+ [2023-03-11 11:50:09,550][00127] Num frames 5800...
1523
+ [2023-03-11 11:50:09,662][00127] Num frames 5900...
1524
+ [2023-03-11 11:50:09,791][00127] Num frames 6000...
1525
+ [2023-03-11 11:50:09,950][00127] Num frames 6100...
1526
+ [2023-03-11 11:50:10,097][00127] Avg episode rewards: #0: 18.653, true rewards: #0: 8.796
1527
+ [2023-03-11 11:50:10,099][00127] Avg episode reward: 18.653, avg true_objective: 8.796
1528
+ [2023-03-11 11:50:10,170][00127] Num frames 6200...
1529
+ [2023-03-11 11:50:10,324][00127] Num frames 6300...
1530
+ [2023-03-11 11:50:10,479][00127] Num frames 6400...
1531
+ [2023-03-11 11:50:10,633][00127] Num frames 6500...
1532
+ [2023-03-11 11:50:10,710][00127] Avg episode rewards: #0: 16.886, true rewards: #0: 8.136
1533
+ [2023-03-11 11:50:10,714][00127] Avg episode reward: 16.886, avg true_objective: 8.136
1534
+ [2023-03-11 11:50:10,861][00127] Num frames 6600...
1535
+ [2023-03-11 11:50:11,011][00127] Num frames 6700...
1536
+ [2023-03-11 11:50:11,161][00127] Num frames 6800...
1537
+ [2023-03-11 11:50:11,320][00127] Num frames 6900...
1538
+ [2023-03-11 11:50:11,473][00127] Num frames 7000...
1539
+ [2023-03-11 11:50:11,634][00127] Num frames 7100...
1540
+ [2023-03-11 11:50:11,794][00127] Num frames 7200...
1541
+ [2023-03-11 11:50:11,969][00127] Num frames 7300...
1542
+ [2023-03-11 11:50:12,154][00127] Num frames 7400...
1543
+ [2023-03-11 11:50:12,328][00127] Num frames 7500...
1544
+ [2023-03-11 11:50:12,499][00127] Num frames 7600...
1545
+ [2023-03-11 11:50:12,663][00127] Num frames 7700...
1546
+ [2023-03-11 11:50:12,823][00127] Num frames 7800...
1547
+ [2023-03-11 11:50:12,980][00127] Num frames 7900...
1548
+ [2023-03-11 11:50:13,146][00127] Num frames 8000...
1549
+ [2023-03-11 11:50:13,323][00127] Avg episode rewards: #0: 19.860, true rewards: #0: 8.971
1550
+ [2023-03-11 11:50:13,325][00127] Avg episode reward: 19.860, avg true_objective: 8.971
1551
+ [2023-03-11 11:50:13,358][00127] Num frames 8100...
1552
+ [2023-03-11 11:50:13,470][00127] Num frames 8200...
1553
+ [2023-03-11 11:50:13,581][00127] Num frames 8300...
1554
+ [2023-03-11 11:50:13,695][00127] Num frames 8400...
1555
+ [2023-03-11 11:50:13,823][00127] Num frames 8500...
1556
+ [2023-03-11 11:50:13,944][00127] Num frames 8600...
1557
+ [2023-03-11 11:50:14,054][00127] Num frames 8700...
1558
+ [2023-03-11 11:50:14,173][00127] Num frames 8800...
1559
+ [2023-03-11 11:50:14,292][00127] Num frames 8900...
1560
+ [2023-03-11 11:50:14,404][00127] Num frames 9000...
1561
+ [2023-03-11 11:50:14,517][00127] Num frames 9100...
1562
+ [2023-03-11 11:50:14,630][00127] Num frames 9200...
1563
+ [2023-03-11 11:50:14,756][00127] Num frames 9300...
1564
+ [2023-03-11 11:50:14,865][00127] Num frames 9400...
1565
+ [2023-03-11 11:50:15,017][00127] Avg episode rewards: #0: 21.182, true rewards: #0: 9.482
1566
+ [2023-03-11 11:50:15,018][00127] Avg episode reward: 21.182, avg true_objective: 9.482
1567
+ [2023-03-11 11:51:10,872][00127] Replay video saved to /content/train_dir/default_experiment/replay.mp4!