Training in progress, step 2240
Browse files- {checkpoint-1920 β checkpoint-2240}/config.json +0 -0
- {checkpoint-1920 β checkpoint-2240}/optimizer.pt +1 -1
- {checkpoint-1920 β checkpoint-2240}/pytorch_model.bin +1 -1
- {checkpoint-1920 β checkpoint-2240}/rng_state.pth +1 -1
- {checkpoint-1920 β checkpoint-2240}/scheduler.pt +1 -1
- {checkpoint-1920 β checkpoint-2240}/trainer_state.json +2195 -3
- {checkpoint-1920 β checkpoint-2240}/training_args.bin +0 -0
- pytorch_model.bin +1 -1
- runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0 +2 -2
{checkpoint-1920 β checkpoint-2240}/config.json
RENAMED
File without changes
|
{checkpoint-1920 β checkpoint-2240}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 677456319
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c222de090e8584f8136998d707c19c56724617254cde9598b5dd6727ac681fe
|
3 |
size 677456319
|
{checkpoint-1920 β checkpoint-2240}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5488ff60e968739c54eef6c3eddadc290534ca260635011c03a8b33733366b1d
|
3 |
size 338796489
|
{checkpoint-1920 β checkpoint-2240}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26ac54f51d16d55e300a23a478923fbe00e997b3bc656cc1b24f0f00a017bcee
|
3 |
size 14575
|
{checkpoint-1920 β checkpoint-2240}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8807874df193d0786d658515555c9c3ee741dfa30a95076b904150645815aa8e
|
3 |
size 627
|
{checkpoint-1920 β checkpoint-2240}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -13158,11 +13158,2203 @@
|
|
13158 |
"eval_samples_per_second": 6.608,
|
13159 |
"eval_steps_per_second": 0.493,
|
13160 |
"step": 1920
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13161 |
}
|
13162 |
],
|
13163 |
"max_steps": 20000,
|
13164 |
"num_train_epochs": 2000,
|
13165 |
-
"total_flos": 1.
|
13166 |
"trial_name": null,
|
13167 |
"trial_params": null
|
13168 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
+
"epoch": 224.0,
|
5 |
+
"global_step": 2240,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
13158 |
"eval_samples_per_second": 6.608,
|
13159 |
"eval_steps_per_second": 0.493,
|
13160 |
"step": 1920
|
13161 |
+
},
|
13162 |
+
{
|
13163 |
+
"epoch": 192.1,
|
13164 |
+
"learning_rate": 9.515263157894737e-05,
|
13165 |
+
"loss": 0.0049,
|
13166 |
+
"step": 1921
|
13167 |
+
},
|
13168 |
+
{
|
13169 |
+
"epoch": 192.2,
|
13170 |
+
"learning_rate": 9.514736842105265e-05,
|
13171 |
+
"loss": 0.0049,
|
13172 |
+
"step": 1922
|
13173 |
+
},
|
13174 |
+
{
|
13175 |
+
"epoch": 192.3,
|
13176 |
+
"learning_rate": 9.51421052631579e-05,
|
13177 |
+
"loss": 0.0053,
|
13178 |
+
"step": 1923
|
13179 |
+
},
|
13180 |
+
{
|
13181 |
+
"epoch": 192.4,
|
13182 |
+
"learning_rate": 9.513684210526316e-05,
|
13183 |
+
"loss": 0.0043,
|
13184 |
+
"step": 1924
|
13185 |
+
},
|
13186 |
+
{
|
13187 |
+
"epoch": 192.5,
|
13188 |
+
"learning_rate": 9.513157894736843e-05,
|
13189 |
+
"loss": 0.0067,
|
13190 |
+
"step": 1925
|
13191 |
+
},
|
13192 |
+
{
|
13193 |
+
"epoch": 192.6,
|
13194 |
+
"learning_rate": 9.512631578947368e-05,
|
13195 |
+
"loss": 0.0045,
|
13196 |
+
"step": 1926
|
13197 |
+
},
|
13198 |
+
{
|
13199 |
+
"epoch": 192.7,
|
13200 |
+
"learning_rate": 9.512105263157896e-05,
|
13201 |
+
"loss": 0.0059,
|
13202 |
+
"step": 1927
|
13203 |
+
},
|
13204 |
+
{
|
13205 |
+
"epoch": 192.8,
|
13206 |
+
"learning_rate": 9.511578947368421e-05,
|
13207 |
+
"loss": 0.0043,
|
13208 |
+
"step": 1928
|
13209 |
+
},
|
13210 |
+
{
|
13211 |
+
"epoch": 192.9,
|
13212 |
+
"learning_rate": 9.511052631578948e-05,
|
13213 |
+
"loss": 0.0048,
|
13214 |
+
"step": 1929
|
13215 |
+
},
|
13216 |
+
{
|
13217 |
+
"epoch": 193.0,
|
13218 |
+
"learning_rate": 9.510526315789475e-05,
|
13219 |
+
"loss": 0.0055,
|
13220 |
+
"step": 1930
|
13221 |
+
},
|
13222 |
+
{
|
13223 |
+
"epoch": 193.1,
|
13224 |
+
"learning_rate": 9.51e-05,
|
13225 |
+
"loss": 0.0046,
|
13226 |
+
"step": 1931
|
13227 |
+
},
|
13228 |
+
{
|
13229 |
+
"epoch": 193.2,
|
13230 |
+
"learning_rate": 9.509473684210527e-05,
|
13231 |
+
"loss": 0.0048,
|
13232 |
+
"step": 1932
|
13233 |
+
},
|
13234 |
+
{
|
13235 |
+
"epoch": 193.3,
|
13236 |
+
"learning_rate": 9.508947368421053e-05,
|
13237 |
+
"loss": 0.0055,
|
13238 |
+
"step": 1933
|
13239 |
+
},
|
13240 |
+
{
|
13241 |
+
"epoch": 193.4,
|
13242 |
+
"learning_rate": 9.50842105263158e-05,
|
13243 |
+
"loss": 0.005,
|
13244 |
+
"step": 1934
|
13245 |
+
},
|
13246 |
+
{
|
13247 |
+
"epoch": 193.5,
|
13248 |
+
"learning_rate": 9.507894736842106e-05,
|
13249 |
+
"loss": 0.0047,
|
13250 |
+
"step": 1935
|
13251 |
+
},
|
13252 |
+
{
|
13253 |
+
"epoch": 193.6,
|
13254 |
+
"learning_rate": 9.507368421052632e-05,
|
13255 |
+
"loss": 0.0052,
|
13256 |
+
"step": 1936
|
13257 |
+
},
|
13258 |
+
{
|
13259 |
+
"epoch": 193.7,
|
13260 |
+
"learning_rate": 9.506842105263158e-05,
|
13261 |
+
"loss": 0.0066,
|
13262 |
+
"step": 1937
|
13263 |
+
},
|
13264 |
+
{
|
13265 |
+
"epoch": 193.8,
|
13266 |
+
"learning_rate": 9.506315789473684e-05,
|
13267 |
+
"loss": 0.0044,
|
13268 |
+
"step": 1938
|
13269 |
+
},
|
13270 |
+
{
|
13271 |
+
"epoch": 193.9,
|
13272 |
+
"learning_rate": 9.50578947368421e-05,
|
13273 |
+
"loss": 0.0054,
|
13274 |
+
"step": 1939
|
13275 |
+
},
|
13276 |
+
{
|
13277 |
+
"epoch": 194.0,
|
13278 |
+
"learning_rate": 9.505263157894737e-05,
|
13279 |
+
"loss": 0.0053,
|
13280 |
+
"step": 1940
|
13281 |
+
},
|
13282 |
+
{
|
13283 |
+
"epoch": 194.0,
|
13284 |
+
"eval_accuracy_safe": 0.843699523132139,
|
13285 |
+
"eval_accuracy_unlabeled": NaN,
|
13286 |
+
"eval_accuracy_unsafe": 0.9961060956490895,
|
13287 |
+
"eval_iou_safe": 0.7480061451912511,
|
13288 |
+
"eval_iou_unlabeled": NaN,
|
13289 |
+
"eval_iou_unsafe": 0.9913896671918935,
|
13290 |
+
"eval_loss": 0.04471457004547119,
|
13291 |
+
"eval_mean_accuracy": 0.9199028093906143,
|
13292 |
+
"eval_mean_iou": 0.8696979061915723,
|
13293 |
+
"eval_overall_accuracy": 0.99160424986882,
|
13294 |
+
"eval_runtime": 10.043,
|
13295 |
+
"eval_samples_per_second": 6.671,
|
13296 |
+
"eval_steps_per_second": 0.498,
|
13297 |
+
"step": 1940
|
13298 |
+
},
|
13299 |
+
{
|
13300 |
+
"epoch": 194.1,
|
13301 |
+
"learning_rate": 9.504736842105264e-05,
|
13302 |
+
"loss": 0.0056,
|
13303 |
+
"step": 1941
|
13304 |
+
},
|
13305 |
+
{
|
13306 |
+
"epoch": 194.2,
|
13307 |
+
"learning_rate": 9.504210526315789e-05,
|
13308 |
+
"loss": 0.0048,
|
13309 |
+
"step": 1942
|
13310 |
+
},
|
13311 |
+
{
|
13312 |
+
"epoch": 194.3,
|
13313 |
+
"learning_rate": 9.503684210526316e-05,
|
13314 |
+
"loss": 0.0041,
|
13315 |
+
"step": 1943
|
13316 |
+
},
|
13317 |
+
{
|
13318 |
+
"epoch": 194.4,
|
13319 |
+
"learning_rate": 9.503157894736843e-05,
|
13320 |
+
"loss": 0.0042,
|
13321 |
+
"step": 1944
|
13322 |
+
},
|
13323 |
+
{
|
13324 |
+
"epoch": 194.5,
|
13325 |
+
"learning_rate": 9.502631578947368e-05,
|
13326 |
+
"loss": 0.0041,
|
13327 |
+
"step": 1945
|
13328 |
+
},
|
13329 |
+
{
|
13330 |
+
"epoch": 194.6,
|
13331 |
+
"learning_rate": 9.502105263157896e-05,
|
13332 |
+
"loss": 0.0058,
|
13333 |
+
"step": 1946
|
13334 |
+
},
|
13335 |
+
{
|
13336 |
+
"epoch": 194.7,
|
13337 |
+
"learning_rate": 9.501578947368422e-05,
|
13338 |
+
"loss": 0.0057,
|
13339 |
+
"step": 1947
|
13340 |
+
},
|
13341 |
+
{
|
13342 |
+
"epoch": 194.8,
|
13343 |
+
"learning_rate": 9.501052631578948e-05,
|
13344 |
+
"loss": 0.004,
|
13345 |
+
"step": 1948
|
13346 |
+
},
|
13347 |
+
{
|
13348 |
+
"epoch": 194.9,
|
13349 |
+
"learning_rate": 9.500526315789474e-05,
|
13350 |
+
"loss": 0.0053,
|
13351 |
+
"step": 1949
|
13352 |
+
},
|
13353 |
+
{
|
13354 |
+
"epoch": 195.0,
|
13355 |
+
"learning_rate": 9.5e-05,
|
13356 |
+
"loss": 0.0061,
|
13357 |
+
"step": 1950
|
13358 |
+
},
|
13359 |
+
{
|
13360 |
+
"epoch": 195.1,
|
13361 |
+
"learning_rate": 9.499473684210527e-05,
|
13362 |
+
"loss": 0.0058,
|
13363 |
+
"step": 1951
|
13364 |
+
},
|
13365 |
+
{
|
13366 |
+
"epoch": 195.2,
|
13367 |
+
"learning_rate": 9.498947368421053e-05,
|
13368 |
+
"loss": 0.0049,
|
13369 |
+
"step": 1952
|
13370 |
+
},
|
13371 |
+
{
|
13372 |
+
"epoch": 195.3,
|
13373 |
+
"learning_rate": 9.498421052631579e-05,
|
13374 |
+
"loss": 0.0055,
|
13375 |
+
"step": 1953
|
13376 |
+
},
|
13377 |
+
{
|
13378 |
+
"epoch": 195.4,
|
13379 |
+
"learning_rate": 9.497894736842105e-05,
|
13380 |
+
"loss": 0.0051,
|
13381 |
+
"step": 1954
|
13382 |
+
},
|
13383 |
+
{
|
13384 |
+
"epoch": 195.5,
|
13385 |
+
"learning_rate": 9.497368421052633e-05,
|
13386 |
+
"loss": 0.005,
|
13387 |
+
"step": 1955
|
13388 |
+
},
|
13389 |
+
{
|
13390 |
+
"epoch": 195.6,
|
13391 |
+
"learning_rate": 9.496842105263158e-05,
|
13392 |
+
"loss": 0.0048,
|
13393 |
+
"step": 1956
|
13394 |
+
},
|
13395 |
+
{
|
13396 |
+
"epoch": 195.7,
|
13397 |
+
"learning_rate": 9.496315789473684e-05,
|
13398 |
+
"loss": 0.0038,
|
13399 |
+
"step": 1957
|
13400 |
+
},
|
13401 |
+
{
|
13402 |
+
"epoch": 195.8,
|
13403 |
+
"learning_rate": 9.495789473684212e-05,
|
13404 |
+
"loss": 0.0041,
|
13405 |
+
"step": 1958
|
13406 |
+
},
|
13407 |
+
{
|
13408 |
+
"epoch": 195.9,
|
13409 |
+
"learning_rate": 9.495263157894737e-05,
|
13410 |
+
"loss": 0.0047,
|
13411 |
+
"step": 1959
|
13412 |
+
},
|
13413 |
+
{
|
13414 |
+
"epoch": 196.0,
|
13415 |
+
"learning_rate": 9.494736842105264e-05,
|
13416 |
+
"loss": 0.0044,
|
13417 |
+
"step": 1960
|
13418 |
+
},
|
13419 |
+
{
|
13420 |
+
"epoch": 196.0,
|
13421 |
+
"eval_accuracy_safe": 0.8516408957559917,
|
13422 |
+
"eval_accuracy_unlabeled": NaN,
|
13423 |
+
"eval_accuracy_unsafe": 0.9958981148905658,
|
13424 |
+
"eval_iou_safe": 0.7505002395034635,
|
13425 |
+
"eval_iou_unlabeled": NaN,
|
13426 |
+
"eval_iou_unsafe": 0.9914211781870088,
|
13427 |
+
"eval_loss": 0.04413814842700958,
|
13428 |
+
"eval_mean_accuracy": 0.9237695053232787,
|
13429 |
+
"eval_mean_iou": 0.8709607088452361,
|
13430 |
+
"eval_overall_accuracy": 0.9916369879423682,
|
13431 |
+
"eval_runtime": 9.6827,
|
13432 |
+
"eval_samples_per_second": 6.92,
|
13433 |
+
"eval_steps_per_second": 0.516,
|
13434 |
+
"step": 1960
|
13435 |
+
},
|
13436 |
+
{
|
13437 |
+
"epoch": 196.1,
|
13438 |
+
"learning_rate": 9.49421052631579e-05,
|
13439 |
+
"loss": 0.0049,
|
13440 |
+
"step": 1961
|
13441 |
+
},
|
13442 |
+
{
|
13443 |
+
"epoch": 196.2,
|
13444 |
+
"learning_rate": 9.493684210526315e-05,
|
13445 |
+
"loss": 0.0055,
|
13446 |
+
"step": 1962
|
13447 |
+
},
|
13448 |
+
{
|
13449 |
+
"epoch": 196.3,
|
13450 |
+
"learning_rate": 9.493157894736843e-05,
|
13451 |
+
"loss": 0.0048,
|
13452 |
+
"step": 1963
|
13453 |
+
},
|
13454 |
+
{
|
13455 |
+
"epoch": 196.4,
|
13456 |
+
"learning_rate": 9.492631578947369e-05,
|
13457 |
+
"loss": 0.004,
|
13458 |
+
"step": 1964
|
13459 |
+
},
|
13460 |
+
{
|
13461 |
+
"epoch": 196.5,
|
13462 |
+
"learning_rate": 9.492105263157895e-05,
|
13463 |
+
"loss": 0.0049,
|
13464 |
+
"step": 1965
|
13465 |
+
},
|
13466 |
+
{
|
13467 |
+
"epoch": 196.6,
|
13468 |
+
"learning_rate": 9.491578947368422e-05,
|
13469 |
+
"loss": 0.0046,
|
13470 |
+
"step": 1966
|
13471 |
+
},
|
13472 |
+
{
|
13473 |
+
"epoch": 196.7,
|
13474 |
+
"learning_rate": 9.491052631578948e-05,
|
13475 |
+
"loss": 0.005,
|
13476 |
+
"step": 1967
|
13477 |
+
},
|
13478 |
+
{
|
13479 |
+
"epoch": 196.8,
|
13480 |
+
"learning_rate": 9.490526315789474e-05,
|
13481 |
+
"loss": 0.0054,
|
13482 |
+
"step": 1968
|
13483 |
+
},
|
13484 |
+
{
|
13485 |
+
"epoch": 196.9,
|
13486 |
+
"learning_rate": 9.49e-05,
|
13487 |
+
"loss": 0.0049,
|
13488 |
+
"step": 1969
|
13489 |
+
},
|
13490 |
+
{
|
13491 |
+
"epoch": 197.0,
|
13492 |
+
"learning_rate": 9.489473684210527e-05,
|
13493 |
+
"loss": 0.0057,
|
13494 |
+
"step": 1970
|
13495 |
+
},
|
13496 |
+
{
|
13497 |
+
"epoch": 197.1,
|
13498 |
+
"learning_rate": 9.488947368421053e-05,
|
13499 |
+
"loss": 0.0041,
|
13500 |
+
"step": 1971
|
13501 |
+
},
|
13502 |
+
{
|
13503 |
+
"epoch": 197.2,
|
13504 |
+
"learning_rate": 9.48842105263158e-05,
|
13505 |
+
"loss": 0.0048,
|
13506 |
+
"step": 1972
|
13507 |
+
},
|
13508 |
+
{
|
13509 |
+
"epoch": 197.3,
|
13510 |
+
"learning_rate": 9.487894736842105e-05,
|
13511 |
+
"loss": 0.0053,
|
13512 |
+
"step": 1973
|
13513 |
+
},
|
13514 |
+
{
|
13515 |
+
"epoch": 197.4,
|
13516 |
+
"learning_rate": 9.487368421052633e-05,
|
13517 |
+
"loss": 0.0052,
|
13518 |
+
"step": 1974
|
13519 |
+
},
|
13520 |
+
{
|
13521 |
+
"epoch": 197.5,
|
13522 |
+
"learning_rate": 9.486842105263159e-05,
|
13523 |
+
"loss": 0.005,
|
13524 |
+
"step": 1975
|
13525 |
+
},
|
13526 |
+
{
|
13527 |
+
"epoch": 197.6,
|
13528 |
+
"learning_rate": 9.486315789473684e-05,
|
13529 |
+
"loss": 0.0048,
|
13530 |
+
"step": 1976
|
13531 |
+
},
|
13532 |
+
{
|
13533 |
+
"epoch": 197.7,
|
13534 |
+
"learning_rate": 9.485789473684211e-05,
|
13535 |
+
"loss": 0.005,
|
13536 |
+
"step": 1977
|
13537 |
+
},
|
13538 |
+
{
|
13539 |
+
"epoch": 197.8,
|
13540 |
+
"learning_rate": 9.485263157894738e-05,
|
13541 |
+
"loss": 0.0043,
|
13542 |
+
"step": 1978
|
13543 |
+
},
|
13544 |
+
{
|
13545 |
+
"epoch": 197.9,
|
13546 |
+
"learning_rate": 9.484736842105264e-05,
|
13547 |
+
"loss": 0.0053,
|
13548 |
+
"step": 1979
|
13549 |
+
},
|
13550 |
+
{
|
13551 |
+
"epoch": 198.0,
|
13552 |
+
"learning_rate": 9.48421052631579e-05,
|
13553 |
+
"loss": 0.0049,
|
13554 |
+
"step": 1980
|
13555 |
+
},
|
13556 |
+
{
|
13557 |
+
"epoch": 198.0,
|
13558 |
+
"eval_accuracy_safe": 0.8479477719823748,
|
13559 |
+
"eval_accuracy_unlabeled": NaN,
|
13560 |
+
"eval_accuracy_unsafe": 0.9959019870288063,
|
13561 |
+
"eval_iou_safe": 0.7473294917880173,
|
13562 |
+
"eval_iou_unlabeled": NaN,
|
13563 |
+
"eval_iou_unsafe": 0.9913141008814026,
|
13564 |
+
"eval_loss": 0.045305173844099045,
|
13565 |
+
"eval_mean_accuracy": 0.9219248795055905,
|
13566 |
+
"eval_mean_iou": 0.86932179633471,
|
13567 |
+
"eval_overall_accuracy": 0.991531656749213,
|
13568 |
+
"eval_runtime": 10.1956,
|
13569 |
+
"eval_samples_per_second": 6.571,
|
13570 |
+
"eval_steps_per_second": 0.49,
|
13571 |
+
"step": 1980
|
13572 |
+
},
|
13573 |
+
{
|
13574 |
+
"epoch": 198.1,
|
13575 |
+
"learning_rate": 9.483684210526315e-05,
|
13576 |
+
"loss": 0.0052,
|
13577 |
+
"step": 1981
|
13578 |
+
},
|
13579 |
+
{
|
13580 |
+
"epoch": 198.2,
|
13581 |
+
"learning_rate": 9.483157894736843e-05,
|
13582 |
+
"loss": 0.005,
|
13583 |
+
"step": 1982
|
13584 |
+
},
|
13585 |
+
{
|
13586 |
+
"epoch": 198.3,
|
13587 |
+
"learning_rate": 9.482631578947369e-05,
|
13588 |
+
"loss": 0.0051,
|
13589 |
+
"step": 1983
|
13590 |
+
},
|
13591 |
+
{
|
13592 |
+
"epoch": 198.4,
|
13593 |
+
"learning_rate": 9.482105263157895e-05,
|
13594 |
+
"loss": 0.0046,
|
13595 |
+
"step": 1984
|
13596 |
+
},
|
13597 |
+
{
|
13598 |
+
"epoch": 198.5,
|
13599 |
+
"learning_rate": 9.481578947368421e-05,
|
13600 |
+
"loss": 0.0049,
|
13601 |
+
"step": 1985
|
13602 |
+
},
|
13603 |
+
{
|
13604 |
+
"epoch": 198.6,
|
13605 |
+
"learning_rate": 9.481052631578949e-05,
|
13606 |
+
"loss": 0.0044,
|
13607 |
+
"step": 1986
|
13608 |
+
},
|
13609 |
+
{
|
13610 |
+
"epoch": 198.7,
|
13611 |
+
"learning_rate": 9.480526315789474e-05,
|
13612 |
+
"loss": 0.0049,
|
13613 |
+
"step": 1987
|
13614 |
+
},
|
13615 |
+
{
|
13616 |
+
"epoch": 198.8,
|
13617 |
+
"learning_rate": 9.48e-05,
|
13618 |
+
"loss": 0.005,
|
13619 |
+
"step": 1988
|
13620 |
+
},
|
13621 |
+
{
|
13622 |
+
"epoch": 198.9,
|
13623 |
+
"learning_rate": 9.479473684210528e-05,
|
13624 |
+
"loss": 0.0045,
|
13625 |
+
"step": 1989
|
13626 |
+
},
|
13627 |
+
{
|
13628 |
+
"epoch": 199.0,
|
13629 |
+
"learning_rate": 9.478947368421053e-05,
|
13630 |
+
"loss": 0.005,
|
13631 |
+
"step": 1990
|
13632 |
+
},
|
13633 |
+
{
|
13634 |
+
"epoch": 199.1,
|
13635 |
+
"learning_rate": 9.47842105263158e-05,
|
13636 |
+
"loss": 0.0043,
|
13637 |
+
"step": 1991
|
13638 |
+
},
|
13639 |
+
{
|
13640 |
+
"epoch": 199.2,
|
13641 |
+
"learning_rate": 9.477894736842105e-05,
|
13642 |
+
"loss": 0.0039,
|
13643 |
+
"step": 1992
|
13644 |
+
},
|
13645 |
+
{
|
13646 |
+
"epoch": 199.3,
|
13647 |
+
"learning_rate": 9.477368421052633e-05,
|
13648 |
+
"loss": 0.0058,
|
13649 |
+
"step": 1993
|
13650 |
+
},
|
13651 |
+
{
|
13652 |
+
"epoch": 199.4,
|
13653 |
+
"learning_rate": 9.476842105263159e-05,
|
13654 |
+
"loss": 0.0042,
|
13655 |
+
"step": 1994
|
13656 |
+
},
|
13657 |
+
{
|
13658 |
+
"epoch": 199.5,
|
13659 |
+
"learning_rate": 9.476315789473684e-05,
|
13660 |
+
"loss": 0.0053,
|
13661 |
+
"step": 1995
|
13662 |
+
},
|
13663 |
+
{
|
13664 |
+
"epoch": 199.6,
|
13665 |
+
"learning_rate": 9.475789473684211e-05,
|
13666 |
+
"loss": 0.005,
|
13667 |
+
"step": 1996
|
13668 |
+
},
|
13669 |
+
{
|
13670 |
+
"epoch": 199.7,
|
13671 |
+
"learning_rate": 9.475263157894737e-05,
|
13672 |
+
"loss": 0.0059,
|
13673 |
+
"step": 1997
|
13674 |
+
},
|
13675 |
+
{
|
13676 |
+
"epoch": 199.8,
|
13677 |
+
"learning_rate": 9.474736842105264e-05,
|
13678 |
+
"loss": 0.0048,
|
13679 |
+
"step": 1998
|
13680 |
+
},
|
13681 |
+
{
|
13682 |
+
"epoch": 199.9,
|
13683 |
+
"learning_rate": 9.47421052631579e-05,
|
13684 |
+
"loss": 0.0043,
|
13685 |
+
"step": 1999
|
13686 |
+
},
|
13687 |
+
{
|
13688 |
+
"epoch": 200.0,
|
13689 |
+
"learning_rate": 9.473684210526316e-05,
|
13690 |
+
"loss": 0.0059,
|
13691 |
+
"step": 2000
|
13692 |
+
},
|
13693 |
+
{
|
13694 |
+
"epoch": 200.0,
|
13695 |
+
"eval_accuracy_safe": 0.8505769060258056,
|
13696 |
+
"eval_accuracy_unlabeled": NaN,
|
13697 |
+
"eval_accuracy_unsafe": 0.9960863242765584,
|
13698 |
+
"eval_iou_safe": 0.7536694505644652,
|
13699 |
+
"eval_iou_unlabeled": NaN,
|
13700 |
+
"eval_iou_unsafe": 0.9915765734959425,
|
13701 |
+
"eval_loss": 0.04439888522028923,
|
13702 |
+
"eval_mean_accuracy": 0.923331615151182,
|
13703 |
+
"eval_mean_iou": 0.8726230120302039,
|
13704 |
+
"eval_overall_accuracy": 0.9917882093742713,
|
13705 |
+
"eval_runtime": 11.3486,
|
13706 |
+
"eval_samples_per_second": 5.904,
|
13707 |
+
"eval_steps_per_second": 0.441,
|
13708 |
+
"step": 2000
|
13709 |
+
},
|
13710 |
+
{
|
13711 |
+
"epoch": 200.1,
|
13712 |
+
"learning_rate": 9.473157894736842e-05,
|
13713 |
+
"loss": 0.0042,
|
13714 |
+
"step": 2001
|
13715 |
+
},
|
13716 |
+
{
|
13717 |
+
"epoch": 200.2,
|
13718 |
+
"learning_rate": 9.472631578947369e-05,
|
13719 |
+
"loss": 0.0055,
|
13720 |
+
"step": 2002
|
13721 |
+
},
|
13722 |
+
{
|
13723 |
+
"epoch": 200.3,
|
13724 |
+
"learning_rate": 9.472105263157895e-05,
|
13725 |
+
"loss": 0.005,
|
13726 |
+
"step": 2003
|
13727 |
+
},
|
13728 |
+
{
|
13729 |
+
"epoch": 200.4,
|
13730 |
+
"learning_rate": 9.471578947368421e-05,
|
13731 |
+
"loss": 0.0049,
|
13732 |
+
"step": 2004
|
13733 |
+
},
|
13734 |
+
{
|
13735 |
+
"epoch": 200.5,
|
13736 |
+
"learning_rate": 9.471052631578949e-05,
|
13737 |
+
"loss": 0.0045,
|
13738 |
+
"step": 2005
|
13739 |
+
},
|
13740 |
+
{
|
13741 |
+
"epoch": 200.6,
|
13742 |
+
"learning_rate": 9.470526315789474e-05,
|
13743 |
+
"loss": 0.0056,
|
13744 |
+
"step": 2006
|
13745 |
+
},
|
13746 |
+
{
|
13747 |
+
"epoch": 200.7,
|
13748 |
+
"learning_rate": 9.47e-05,
|
13749 |
+
"loss": 0.0043,
|
13750 |
+
"step": 2007
|
13751 |
+
},
|
13752 |
+
{
|
13753 |
+
"epoch": 200.8,
|
13754 |
+
"learning_rate": 9.469473684210527e-05,
|
13755 |
+
"loss": 0.0048,
|
13756 |
+
"step": 2008
|
13757 |
+
},
|
13758 |
+
{
|
13759 |
+
"epoch": 200.9,
|
13760 |
+
"learning_rate": 9.468947368421052e-05,
|
13761 |
+
"loss": 0.0048,
|
13762 |
+
"step": 2009
|
13763 |
+
},
|
13764 |
+
{
|
13765 |
+
"epoch": 201.0,
|
13766 |
+
"learning_rate": 9.46842105263158e-05,
|
13767 |
+
"loss": 0.0041,
|
13768 |
+
"step": 2010
|
13769 |
+
},
|
13770 |
+
{
|
13771 |
+
"epoch": 201.1,
|
13772 |
+
"learning_rate": 9.467894736842106e-05,
|
13773 |
+
"loss": 0.0049,
|
13774 |
+
"step": 2011
|
13775 |
+
},
|
13776 |
+
{
|
13777 |
+
"epoch": 201.2,
|
13778 |
+
"learning_rate": 9.467368421052632e-05,
|
13779 |
+
"loss": 0.0049,
|
13780 |
+
"step": 2012
|
13781 |
+
},
|
13782 |
+
{
|
13783 |
+
"epoch": 201.3,
|
13784 |
+
"learning_rate": 9.466842105263159e-05,
|
13785 |
+
"loss": 0.0038,
|
13786 |
+
"step": 2013
|
13787 |
+
},
|
13788 |
+
{
|
13789 |
+
"epoch": 201.4,
|
13790 |
+
"learning_rate": 9.466315789473685e-05,
|
13791 |
+
"loss": 0.0046,
|
13792 |
+
"step": 2014
|
13793 |
+
},
|
13794 |
+
{
|
13795 |
+
"epoch": 201.5,
|
13796 |
+
"learning_rate": 9.465789473684211e-05,
|
13797 |
+
"loss": 0.0044,
|
13798 |
+
"step": 2015
|
13799 |
+
},
|
13800 |
+
{
|
13801 |
+
"epoch": 201.6,
|
13802 |
+
"learning_rate": 9.465263157894737e-05,
|
13803 |
+
"loss": 0.0041,
|
13804 |
+
"step": 2016
|
13805 |
+
},
|
13806 |
+
{
|
13807 |
+
"epoch": 201.7,
|
13808 |
+
"learning_rate": 9.464736842105263e-05,
|
13809 |
+
"loss": 0.0058,
|
13810 |
+
"step": 2017
|
13811 |
+
},
|
13812 |
+
{
|
13813 |
+
"epoch": 201.8,
|
13814 |
+
"learning_rate": 9.46421052631579e-05,
|
13815 |
+
"loss": 0.005,
|
13816 |
+
"step": 2018
|
13817 |
+
},
|
13818 |
+
{
|
13819 |
+
"epoch": 201.9,
|
13820 |
+
"learning_rate": 9.463684210526316e-05,
|
13821 |
+
"loss": 0.0052,
|
13822 |
+
"step": 2019
|
13823 |
+
},
|
13824 |
+
{
|
13825 |
+
"epoch": 202.0,
|
13826 |
+
"learning_rate": 9.463157894736842e-05,
|
13827 |
+
"loss": 0.005,
|
13828 |
+
"step": 2020
|
13829 |
+
},
|
13830 |
+
{
|
13831 |
+
"epoch": 202.0,
|
13832 |
+
"eval_accuracy_safe": 0.8554862934221533,
|
13833 |
+
"eval_accuracy_unlabeled": NaN,
|
13834 |
+
"eval_accuracy_unsafe": 0.9958080583420935,
|
13835 |
+
"eval_iou_safe": 0.7519284103596254,
|
13836 |
+
"eval_iou_unlabeled": NaN,
|
13837 |
+
"eval_iou_unsafe": 0.9914470479245693,
|
13838 |
+
"eval_loss": 0.04469138756394386,
|
13839 |
+
"eval_mean_accuracy": 0.9256471758821234,
|
13840 |
+
"eval_mean_iou": 0.8716877291420974,
|
13841 |
+
"eval_overall_accuracy": 0.9916631784012069,
|
13842 |
+
"eval_runtime": 10.1291,
|
13843 |
+
"eval_samples_per_second": 6.615,
|
13844 |
+
"eval_steps_per_second": 0.494,
|
13845 |
+
"step": 2020
|
13846 |
+
},
|
13847 |
+
{
|
13848 |
+
"epoch": 202.1,
|
13849 |
+
"learning_rate": 9.462631578947368e-05,
|
13850 |
+
"loss": 0.0042,
|
13851 |
+
"step": 2021
|
13852 |
+
},
|
13853 |
+
{
|
13854 |
+
"epoch": 202.2,
|
13855 |
+
"learning_rate": 9.462105263157896e-05,
|
13856 |
+
"loss": 0.0046,
|
13857 |
+
"step": 2022
|
13858 |
+
},
|
13859 |
+
{
|
13860 |
+
"epoch": 202.3,
|
13861 |
+
"learning_rate": 9.461578947368421e-05,
|
13862 |
+
"loss": 0.0049,
|
13863 |
+
"step": 2023
|
13864 |
+
},
|
13865 |
+
{
|
13866 |
+
"epoch": 202.4,
|
13867 |
+
"learning_rate": 9.461052631578948e-05,
|
13868 |
+
"loss": 0.0052,
|
13869 |
+
"step": 2024
|
13870 |
+
},
|
13871 |
+
{
|
13872 |
+
"epoch": 202.5,
|
13873 |
+
"learning_rate": 9.460526315789475e-05,
|
13874 |
+
"loss": 0.0051,
|
13875 |
+
"step": 2025
|
13876 |
+
},
|
13877 |
+
{
|
13878 |
+
"epoch": 202.6,
|
13879 |
+
"learning_rate": 9.46e-05,
|
13880 |
+
"loss": 0.0046,
|
13881 |
+
"step": 2026
|
13882 |
+
},
|
13883 |
+
{
|
13884 |
+
"epoch": 202.7,
|
13885 |
+
"learning_rate": 9.459473684210527e-05,
|
13886 |
+
"loss": 0.0045,
|
13887 |
+
"step": 2027
|
13888 |
+
},
|
13889 |
+
{
|
13890 |
+
"epoch": 202.8,
|
13891 |
+
"learning_rate": 9.458947368421053e-05,
|
13892 |
+
"loss": 0.0043,
|
13893 |
+
"step": 2028
|
13894 |
+
},
|
13895 |
+
{
|
13896 |
+
"epoch": 202.9,
|
13897 |
+
"learning_rate": 9.45842105263158e-05,
|
13898 |
+
"loss": 0.0045,
|
13899 |
+
"step": 2029
|
13900 |
+
},
|
13901 |
+
{
|
13902 |
+
"epoch": 203.0,
|
13903 |
+
"learning_rate": 9.457894736842106e-05,
|
13904 |
+
"loss": 0.0057,
|
13905 |
+
"step": 2030
|
13906 |
+
},
|
13907 |
+
{
|
13908 |
+
"epoch": 203.1,
|
13909 |
+
"learning_rate": 9.457368421052632e-05,
|
13910 |
+
"loss": 0.0053,
|
13911 |
+
"step": 2031
|
13912 |
+
},
|
13913 |
+
{
|
13914 |
+
"epoch": 203.2,
|
13915 |
+
"learning_rate": 9.456842105263158e-05,
|
13916 |
+
"loss": 0.0045,
|
13917 |
+
"step": 2032
|
13918 |
+
},
|
13919 |
+
{
|
13920 |
+
"epoch": 203.3,
|
13921 |
+
"learning_rate": 9.456315789473685e-05,
|
13922 |
+
"loss": 0.0045,
|
13923 |
+
"step": 2033
|
13924 |
+
},
|
13925 |
+
{
|
13926 |
+
"epoch": 203.4,
|
13927 |
+
"learning_rate": 9.455789473684211e-05,
|
13928 |
+
"loss": 0.0047,
|
13929 |
+
"step": 2034
|
13930 |
+
},
|
13931 |
+
{
|
13932 |
+
"epoch": 203.5,
|
13933 |
+
"learning_rate": 9.455263157894737e-05,
|
13934 |
+
"loss": 0.0053,
|
13935 |
+
"step": 2035
|
13936 |
+
},
|
13937 |
+
{
|
13938 |
+
"epoch": 203.6,
|
13939 |
+
"learning_rate": 9.454736842105265e-05,
|
13940 |
+
"loss": 0.0044,
|
13941 |
+
"step": 2036
|
13942 |
+
},
|
13943 |
+
{
|
13944 |
+
"epoch": 203.7,
|
13945 |
+
"learning_rate": 9.45421052631579e-05,
|
13946 |
+
"loss": 0.0057,
|
13947 |
+
"step": 2037
|
13948 |
+
},
|
13949 |
+
{
|
13950 |
+
"epoch": 203.8,
|
13951 |
+
"learning_rate": 9.453684210526316e-05,
|
13952 |
+
"loss": 0.0049,
|
13953 |
+
"step": 2038
|
13954 |
+
},
|
13955 |
+
{
|
13956 |
+
"epoch": 203.9,
|
13957 |
+
"learning_rate": 9.453157894736843e-05,
|
13958 |
+
"loss": 0.0041,
|
13959 |
+
"step": 2039
|
13960 |
+
},
|
13961 |
+
{
|
13962 |
+
"epoch": 204.0,
|
13963 |
+
"learning_rate": 9.452631578947368e-05,
|
13964 |
+
"loss": 0.005,
|
13965 |
+
"step": 2040
|
13966 |
+
},
|
13967 |
+
{
|
13968 |
+
"epoch": 204.0,
|
13969 |
+
"eval_accuracy_safe": 0.8493991927556177,
|
13970 |
+
"eval_accuracy_unlabeled": NaN,
|
13971 |
+
"eval_accuracy_unsafe": 0.9959986731472963,
|
13972 |
+
"eval_iou_safe": 0.7507103869820307,
|
13973 |
+
"eval_iou_unlabeled": NaN,
|
13974 |
+
"eval_iou_unsafe": 0.9914539398462113,
|
13975 |
+
"eval_loss": 0.04513503238558769,
|
13976 |
+
"eval_mean_accuracy": 0.922698932951457,
|
13977 |
+
"eval_mean_iou": 0.871082163414121,
|
13978 |
+
"eval_overall_accuracy": 0.9916683595571945,
|
13979 |
+
"eval_runtime": 9.8572,
|
13980 |
+
"eval_samples_per_second": 6.797,
|
13981 |
+
"eval_steps_per_second": 0.507,
|
13982 |
+
"step": 2040
|
13983 |
+
},
|
13984 |
+
{
|
13985 |
+
"epoch": 204.1,
|
13986 |
+
"learning_rate": 9.452105263157896e-05,
|
13987 |
+
"loss": 0.0045,
|
13988 |
+
"step": 2041
|
13989 |
+
},
|
13990 |
+
{
|
13991 |
+
"epoch": 204.2,
|
13992 |
+
"learning_rate": 9.451578947368422e-05,
|
13993 |
+
"loss": 0.0046,
|
13994 |
+
"step": 2042
|
13995 |
+
},
|
13996 |
+
{
|
13997 |
+
"epoch": 204.3,
|
13998 |
+
"learning_rate": 9.451052631578948e-05,
|
13999 |
+
"loss": 0.005,
|
14000 |
+
"step": 2043
|
14001 |
+
},
|
14002 |
+
{
|
14003 |
+
"epoch": 204.4,
|
14004 |
+
"learning_rate": 9.450526315789474e-05,
|
14005 |
+
"loss": 0.0057,
|
14006 |
+
"step": 2044
|
14007 |
+
},
|
14008 |
+
{
|
14009 |
+
"epoch": 204.5,
|
14010 |
+
"learning_rate": 9.449999999999999e-05,
|
14011 |
+
"loss": 0.0041,
|
14012 |
+
"step": 2045
|
14013 |
+
},
|
14014 |
+
{
|
14015 |
+
"epoch": 204.6,
|
14016 |
+
"learning_rate": 9.449473684210527e-05,
|
14017 |
+
"loss": 0.0051,
|
14018 |
+
"step": 2046
|
14019 |
+
},
|
14020 |
+
{
|
14021 |
+
"epoch": 204.7,
|
14022 |
+
"learning_rate": 9.448947368421053e-05,
|
14023 |
+
"loss": 0.0051,
|
14024 |
+
"step": 2047
|
14025 |
+
},
|
14026 |
+
{
|
14027 |
+
"epoch": 204.8,
|
14028 |
+
"learning_rate": 9.44842105263158e-05,
|
14029 |
+
"loss": 0.0049,
|
14030 |
+
"step": 2048
|
14031 |
+
},
|
14032 |
+
{
|
14033 |
+
"epoch": 204.9,
|
14034 |
+
"learning_rate": 9.447894736842106e-05,
|
14035 |
+
"loss": 0.0043,
|
14036 |
+
"step": 2049
|
14037 |
+
},
|
14038 |
+
{
|
14039 |
+
"epoch": 205.0,
|
14040 |
+
"learning_rate": 9.447368421052633e-05,
|
14041 |
+
"loss": 0.0048,
|
14042 |
+
"step": 2050
|
14043 |
+
},
|
14044 |
+
{
|
14045 |
+
"epoch": 205.1,
|
14046 |
+
"learning_rate": 9.446842105263158e-05,
|
14047 |
+
"loss": 0.0038,
|
14048 |
+
"step": 2051
|
14049 |
+
},
|
14050 |
+
{
|
14051 |
+
"epoch": 205.2,
|
14052 |
+
"learning_rate": 9.446315789473684e-05,
|
14053 |
+
"loss": 0.0044,
|
14054 |
+
"step": 2052
|
14055 |
+
},
|
14056 |
+
{
|
14057 |
+
"epoch": 205.3,
|
14058 |
+
"learning_rate": 9.44578947368421e-05,
|
14059 |
+
"loss": 0.0056,
|
14060 |
+
"step": 2053
|
14061 |
+
},
|
14062 |
+
{
|
14063 |
+
"epoch": 205.4,
|
14064 |
+
"learning_rate": 9.445263157894737e-05,
|
14065 |
+
"loss": 0.0058,
|
14066 |
+
"step": 2054
|
14067 |
+
},
|
14068 |
+
{
|
14069 |
+
"epoch": 205.5,
|
14070 |
+
"learning_rate": 9.444736842105264e-05,
|
14071 |
+
"loss": 0.0046,
|
14072 |
+
"step": 2055
|
14073 |
+
},
|
14074 |
+
{
|
14075 |
+
"epoch": 205.6,
|
14076 |
+
"learning_rate": 9.444210526315789e-05,
|
14077 |
+
"loss": 0.0051,
|
14078 |
+
"step": 2056
|
14079 |
+
},
|
14080 |
+
{
|
14081 |
+
"epoch": 205.7,
|
14082 |
+
"learning_rate": 9.443684210526315e-05,
|
14083 |
+
"loss": 0.0042,
|
14084 |
+
"step": 2057
|
14085 |
+
},
|
14086 |
+
{
|
14087 |
+
"epoch": 205.8,
|
14088 |
+
"learning_rate": 9.443157894736843e-05,
|
14089 |
+
"loss": 0.0045,
|
14090 |
+
"step": 2058
|
14091 |
+
},
|
14092 |
+
{
|
14093 |
+
"epoch": 205.9,
|
14094 |
+
"learning_rate": 9.442631578947368e-05,
|
14095 |
+
"loss": 0.0044,
|
14096 |
+
"step": 2059
|
14097 |
+
},
|
14098 |
+
{
|
14099 |
+
"epoch": 206.0,
|
14100 |
+
"learning_rate": 9.442105263157895e-05,
|
14101 |
+
"loss": 0.0043,
|
14102 |
+
"step": 2060
|
14103 |
+
},
|
14104 |
+
{
|
14105 |
+
"epoch": 206.0,
|
14106 |
+
"eval_accuracy_safe": 0.8479554820528834,
|
14107 |
+
"eval_accuracy_unlabeled": NaN,
|
14108 |
+
"eval_accuracy_unsafe": 0.9960219059767392,
|
14109 |
+
"eval_iou_safe": 0.7499403349425852,
|
14110 |
+
"eval_iou_unlabeled": NaN,
|
14111 |
+
"eval_iou_unsafe": 0.9914336989838719,
|
14112 |
+
"eval_loss": 0.04577173292636871,
|
14113 |
+
"eval_mean_accuracy": 0.9219886940148113,
|
14114 |
+
"eval_mean_iou": 0.8706870169632286,
|
14115 |
+
"eval_overall_accuracy": 0.9916482612268248,
|
14116 |
+
"eval_runtime": 9.6732,
|
14117 |
+
"eval_samples_per_second": 6.926,
|
14118 |
+
"eval_steps_per_second": 0.517,
|
14119 |
+
"step": 2060
|
14120 |
+
},
|
14121 |
+
{
|
14122 |
+
"epoch": 206.1,
|
14123 |
+
"learning_rate": 9.441578947368422e-05,
|
14124 |
+
"loss": 0.0046,
|
14125 |
+
"step": 2061
|
14126 |
+
},
|
14127 |
+
{
|
14128 |
+
"epoch": 206.2,
|
14129 |
+
"learning_rate": 9.441052631578948e-05,
|
14130 |
+
"loss": 0.0048,
|
14131 |
+
"step": 2062
|
14132 |
+
},
|
14133 |
+
{
|
14134 |
+
"epoch": 206.3,
|
14135 |
+
"learning_rate": 9.440526315789474e-05,
|
14136 |
+
"loss": 0.0043,
|
14137 |
+
"step": 2063
|
14138 |
+
},
|
14139 |
+
{
|
14140 |
+
"epoch": 206.4,
|
14141 |
+
"learning_rate": 9.44e-05,
|
14142 |
+
"loss": 0.0044,
|
14143 |
+
"step": 2064
|
14144 |
+
},
|
14145 |
+
{
|
14146 |
+
"epoch": 206.5,
|
14147 |
+
"learning_rate": 9.439473684210527e-05,
|
14148 |
+
"loss": 0.0051,
|
14149 |
+
"step": 2065
|
14150 |
+
},
|
14151 |
+
{
|
14152 |
+
"epoch": 206.6,
|
14153 |
+
"learning_rate": 9.438947368421053e-05,
|
14154 |
+
"loss": 0.005,
|
14155 |
+
"step": 2066
|
14156 |
+
},
|
14157 |
+
{
|
14158 |
+
"epoch": 206.7,
|
14159 |
+
"learning_rate": 9.438421052631579e-05,
|
14160 |
+
"loss": 0.0051,
|
14161 |
+
"step": 2067
|
14162 |
+
},
|
14163 |
+
{
|
14164 |
+
"epoch": 206.8,
|
14165 |
+
"learning_rate": 9.437894736842105e-05,
|
14166 |
+
"loss": 0.0043,
|
14167 |
+
"step": 2068
|
14168 |
+
},
|
14169 |
+
{
|
14170 |
+
"epoch": 206.9,
|
14171 |
+
"learning_rate": 9.437368421052633e-05,
|
14172 |
+
"loss": 0.0048,
|
14173 |
+
"step": 2069
|
14174 |
+
},
|
14175 |
+
{
|
14176 |
+
"epoch": 207.0,
|
14177 |
+
"learning_rate": 9.436842105263158e-05,
|
14178 |
+
"loss": 0.0052,
|
14179 |
+
"step": 2070
|
14180 |
+
},
|
14181 |
+
{
|
14182 |
+
"epoch": 207.1,
|
14183 |
+
"learning_rate": 9.436315789473684e-05,
|
14184 |
+
"loss": 0.0045,
|
14185 |
+
"step": 2071
|
14186 |
+
},
|
14187 |
+
{
|
14188 |
+
"epoch": 207.2,
|
14189 |
+
"learning_rate": 9.435789473684212e-05,
|
14190 |
+
"loss": 0.0044,
|
14191 |
+
"step": 2072
|
14192 |
+
},
|
14193 |
+
{
|
14194 |
+
"epoch": 207.3,
|
14195 |
+
"learning_rate": 9.435263157894737e-05,
|
14196 |
+
"loss": 0.0047,
|
14197 |
+
"step": 2073
|
14198 |
+
},
|
14199 |
+
{
|
14200 |
+
"epoch": 207.4,
|
14201 |
+
"learning_rate": 9.434736842105264e-05,
|
14202 |
+
"loss": 0.005,
|
14203 |
+
"step": 2074
|
14204 |
+
},
|
14205 |
+
{
|
14206 |
+
"epoch": 207.5,
|
14207 |
+
"learning_rate": 9.43421052631579e-05,
|
14208 |
+
"loss": 0.0047,
|
14209 |
+
"step": 2075
|
14210 |
+
},
|
14211 |
+
{
|
14212 |
+
"epoch": 207.6,
|
14213 |
+
"learning_rate": 9.433684210526317e-05,
|
14214 |
+
"loss": 0.0057,
|
14215 |
+
"step": 2076
|
14216 |
+
},
|
14217 |
+
{
|
14218 |
+
"epoch": 207.7,
|
14219 |
+
"learning_rate": 9.433157894736843e-05,
|
14220 |
+
"loss": 0.0045,
|
14221 |
+
"step": 2077
|
14222 |
+
},
|
14223 |
+
{
|
14224 |
+
"epoch": 207.8,
|
14225 |
+
"learning_rate": 9.432631578947369e-05,
|
14226 |
+
"loss": 0.0043,
|
14227 |
+
"step": 2078
|
14228 |
+
},
|
14229 |
+
{
|
14230 |
+
"epoch": 207.9,
|
14231 |
+
"learning_rate": 9.432105263157895e-05,
|
14232 |
+
"loss": 0.0049,
|
14233 |
+
"step": 2079
|
14234 |
+
},
|
14235 |
+
{
|
14236 |
+
"epoch": 208.0,
|
14237 |
+
"learning_rate": 9.431578947368421e-05,
|
14238 |
+
"loss": 0.0043,
|
14239 |
+
"step": 2080
|
14240 |
+
},
|
14241 |
+
{
|
14242 |
+
"epoch": 208.0,
|
14243 |
+
"eval_accuracy_safe": 0.8482195519678027,
|
14244 |
+
"eval_accuracy_unlabeled": NaN,
|
14245 |
+
"eval_accuracy_unsafe": 0.9959190596383212,
|
14246 |
+
"eval_iou_safe": 0.7479387654858157,
|
14247 |
+
"eval_iou_unlabeled": NaN,
|
14248 |
+
"eval_iou_unsafe": 0.9913392577140658,
|
14249 |
+
"eval_loss": 0.04617840424180031,
|
14250 |
+
"eval_mean_accuracy": 0.922069305803062,
|
14251 |
+
"eval_mean_iou": 0.8696390115999408,
|
14252 |
+
"eval_overall_accuracy": 0.9915562530062092,
|
14253 |
+
"eval_runtime": 9.8887,
|
14254 |
+
"eval_samples_per_second": 6.775,
|
14255 |
+
"eval_steps_per_second": 0.506,
|
14256 |
+
"step": 2080
|
14257 |
+
},
|
14258 |
+
{
|
14259 |
+
"epoch": 208.1,
|
14260 |
+
"learning_rate": 9.431052631578948e-05,
|
14261 |
+
"loss": 0.0049,
|
14262 |
+
"step": 2081
|
14263 |
+
},
|
14264 |
+
{
|
14265 |
+
"epoch": 208.2,
|
14266 |
+
"learning_rate": 9.430526315789474e-05,
|
14267 |
+
"loss": 0.0043,
|
14268 |
+
"step": 2082
|
14269 |
+
},
|
14270 |
+
{
|
14271 |
+
"epoch": 208.3,
|
14272 |
+
"learning_rate": 9.43e-05,
|
14273 |
+
"loss": 0.0044,
|
14274 |
+
"step": 2083
|
14275 |
+
},
|
14276 |
+
{
|
14277 |
+
"epoch": 208.4,
|
14278 |
+
"learning_rate": 9.429473684210526e-05,
|
14279 |
+
"loss": 0.0044,
|
14280 |
+
"step": 2084
|
14281 |
+
},
|
14282 |
+
{
|
14283 |
+
"epoch": 208.5,
|
14284 |
+
"learning_rate": 9.428947368421053e-05,
|
14285 |
+
"loss": 0.0052,
|
14286 |
+
"step": 2085
|
14287 |
+
},
|
14288 |
+
{
|
14289 |
+
"epoch": 208.6,
|
14290 |
+
"learning_rate": 9.42842105263158e-05,
|
14291 |
+
"loss": 0.0048,
|
14292 |
+
"step": 2086
|
14293 |
+
},
|
14294 |
+
{
|
14295 |
+
"epoch": 208.7,
|
14296 |
+
"learning_rate": 9.427894736842105e-05,
|
14297 |
+
"loss": 0.0041,
|
14298 |
+
"step": 2087
|
14299 |
+
},
|
14300 |
+
{
|
14301 |
+
"epoch": 208.8,
|
14302 |
+
"learning_rate": 9.427368421052633e-05,
|
14303 |
+
"loss": 0.0049,
|
14304 |
+
"step": 2088
|
14305 |
+
},
|
14306 |
+
{
|
14307 |
+
"epoch": 208.9,
|
14308 |
+
"learning_rate": 9.426842105263159e-05,
|
14309 |
+
"loss": 0.0049,
|
14310 |
+
"step": 2089
|
14311 |
+
},
|
14312 |
+
{
|
14313 |
+
"epoch": 209.0,
|
14314 |
+
"learning_rate": 9.426315789473684e-05,
|
14315 |
+
"loss": 0.0053,
|
14316 |
+
"step": 2090
|
14317 |
+
},
|
14318 |
+
{
|
14319 |
+
"epoch": 209.1,
|
14320 |
+
"learning_rate": 9.425789473684211e-05,
|
14321 |
+
"loss": 0.0041,
|
14322 |
+
"step": 2091
|
14323 |
+
},
|
14324 |
+
{
|
14325 |
+
"epoch": 209.2,
|
14326 |
+
"learning_rate": 9.425263157894738e-05,
|
14327 |
+
"loss": 0.0036,
|
14328 |
+
"step": 2092
|
14329 |
+
},
|
14330 |
+
{
|
14331 |
+
"epoch": 209.3,
|
14332 |
+
"learning_rate": 9.424736842105264e-05,
|
14333 |
+
"loss": 0.0042,
|
14334 |
+
"step": 2093
|
14335 |
+
},
|
14336 |
+
{
|
14337 |
+
"epoch": 209.4,
|
14338 |
+
"learning_rate": 9.42421052631579e-05,
|
14339 |
+
"loss": 0.0048,
|
14340 |
+
"step": 2094
|
14341 |
+
},
|
14342 |
+
{
|
14343 |
+
"epoch": 209.5,
|
14344 |
+
"learning_rate": 9.423684210526316e-05,
|
14345 |
+
"loss": 0.0053,
|
14346 |
+
"step": 2095
|
14347 |
+
},
|
14348 |
+
{
|
14349 |
+
"epoch": 209.6,
|
14350 |
+
"learning_rate": 9.423157894736843e-05,
|
14351 |
+
"loss": 0.0052,
|
14352 |
+
"step": 2096
|
14353 |
+
},
|
14354 |
+
{
|
14355 |
+
"epoch": 209.7,
|
14356 |
+
"learning_rate": 9.422631578947369e-05,
|
14357 |
+
"loss": 0.0042,
|
14358 |
+
"step": 2097
|
14359 |
+
},
|
14360 |
+
{
|
14361 |
+
"epoch": 209.8,
|
14362 |
+
"learning_rate": 9.422105263157895e-05,
|
14363 |
+
"loss": 0.0043,
|
14364 |
+
"step": 2098
|
14365 |
+
},
|
14366 |
+
{
|
14367 |
+
"epoch": 209.9,
|
14368 |
+
"learning_rate": 9.421578947368421e-05,
|
14369 |
+
"loss": 0.0051,
|
14370 |
+
"step": 2099
|
14371 |
+
},
|
14372 |
+
{
|
14373 |
+
"epoch": 210.0,
|
14374 |
+
"learning_rate": 9.421052631578949e-05,
|
14375 |
+
"loss": 0.0062,
|
14376 |
+
"step": 2100
|
14377 |
+
},
|
14378 |
+
{
|
14379 |
+
"epoch": 210.0,
|
14380 |
+
"eval_accuracy_safe": 0.846122412789465,
|
14381 |
+
"eval_accuracy_unlabeled": NaN,
|
14382 |
+
"eval_accuracy_unsafe": 0.9961696925862515,
|
14383 |
+
"eval_iou_safe": 0.751546425287927,
|
14384 |
+
"eval_iou_unlabeled": NaN,
|
14385 |
+
"eval_iou_unsafe": 0.9915257384693599,
|
14386 |
+
"eval_loss": 0.04509461671113968,
|
14387 |
+
"eval_mean_accuracy": 0.9211460526878583,
|
14388 |
+
"eval_mean_iou": 0.8715360818786435,
|
14389 |
+
"eval_overall_accuracy": 0.9917375365299965,
|
14390 |
+
"eval_runtime": 9.5411,
|
14391 |
+
"eval_samples_per_second": 7.022,
|
14392 |
+
"eval_steps_per_second": 0.524,
|
14393 |
+
"step": 2100
|
14394 |
+
},
|
14395 |
+
{
|
14396 |
+
"epoch": 210.1,
|
14397 |
+
"learning_rate": 9.420526315789474e-05,
|
14398 |
+
"loss": 0.0043,
|
14399 |
+
"step": 2101
|
14400 |
+
},
|
14401 |
+
{
|
14402 |
+
"epoch": 210.2,
|
14403 |
+
"learning_rate": 9.42e-05,
|
14404 |
+
"loss": 0.0054,
|
14405 |
+
"step": 2102
|
14406 |
+
},
|
14407 |
+
{
|
14408 |
+
"epoch": 210.3,
|
14409 |
+
"learning_rate": 9.419473684210528e-05,
|
14410 |
+
"loss": 0.0051,
|
14411 |
+
"step": 2103
|
14412 |
+
},
|
14413 |
+
{
|
14414 |
+
"epoch": 210.4,
|
14415 |
+
"learning_rate": 9.418947368421052e-05,
|
14416 |
+
"loss": 0.0041,
|
14417 |
+
"step": 2104
|
14418 |
+
},
|
14419 |
+
{
|
14420 |
+
"epoch": 210.5,
|
14421 |
+
"learning_rate": 9.41842105263158e-05,
|
14422 |
+
"loss": 0.0046,
|
14423 |
+
"step": 2105
|
14424 |
+
},
|
14425 |
+
{
|
14426 |
+
"epoch": 210.6,
|
14427 |
+
"learning_rate": 9.417894736842106e-05,
|
14428 |
+
"loss": 0.0049,
|
14429 |
+
"step": 2106
|
14430 |
+
},
|
14431 |
+
{
|
14432 |
+
"epoch": 210.7,
|
14433 |
+
"learning_rate": 9.417368421052632e-05,
|
14434 |
+
"loss": 0.0042,
|
14435 |
+
"step": 2107
|
14436 |
+
},
|
14437 |
+
{
|
14438 |
+
"epoch": 210.8,
|
14439 |
+
"learning_rate": 9.416842105263159e-05,
|
14440 |
+
"loss": 0.0047,
|
14441 |
+
"step": 2108
|
14442 |
+
},
|
14443 |
+
{
|
14444 |
+
"epoch": 210.9,
|
14445 |
+
"learning_rate": 9.416315789473684e-05,
|
14446 |
+
"loss": 0.0051,
|
14447 |
+
"step": 2109
|
14448 |
+
},
|
14449 |
+
{
|
14450 |
+
"epoch": 211.0,
|
14451 |
+
"learning_rate": 9.415789473684211e-05,
|
14452 |
+
"loss": 0.0045,
|
14453 |
+
"step": 2110
|
14454 |
+
},
|
14455 |
+
{
|
14456 |
+
"epoch": 211.1,
|
14457 |
+
"learning_rate": 9.415263157894737e-05,
|
14458 |
+
"loss": 0.0039,
|
14459 |
+
"step": 2111
|
14460 |
+
},
|
14461 |
+
{
|
14462 |
+
"epoch": 211.2,
|
14463 |
+
"learning_rate": 9.414736842105264e-05,
|
14464 |
+
"loss": 0.0047,
|
14465 |
+
"step": 2112
|
14466 |
+
},
|
14467 |
+
{
|
14468 |
+
"epoch": 211.3,
|
14469 |
+
"learning_rate": 9.41421052631579e-05,
|
14470 |
+
"loss": 0.0049,
|
14471 |
+
"step": 2113
|
14472 |
+
},
|
14473 |
+
{
|
14474 |
+
"epoch": 211.4,
|
14475 |
+
"learning_rate": 9.413684210526317e-05,
|
14476 |
+
"loss": 0.0046,
|
14477 |
+
"step": 2114
|
14478 |
+
},
|
14479 |
+
{
|
14480 |
+
"epoch": 211.5,
|
14481 |
+
"learning_rate": 9.413157894736842e-05,
|
14482 |
+
"loss": 0.0042,
|
14483 |
+
"step": 2115
|
14484 |
+
},
|
14485 |
+
{
|
14486 |
+
"epoch": 211.6,
|
14487 |
+
"learning_rate": 9.412631578947369e-05,
|
14488 |
+
"loss": 0.0045,
|
14489 |
+
"step": 2116
|
14490 |
+
},
|
14491 |
+
{
|
14492 |
+
"epoch": 211.7,
|
14493 |
+
"learning_rate": 9.412105263157895e-05,
|
14494 |
+
"loss": 0.0049,
|
14495 |
+
"step": 2117
|
14496 |
+
},
|
14497 |
+
{
|
14498 |
+
"epoch": 211.8,
|
14499 |
+
"learning_rate": 9.411578947368421e-05,
|
14500 |
+
"loss": 0.0053,
|
14501 |
+
"step": 2118
|
14502 |
+
},
|
14503 |
+
{
|
14504 |
+
"epoch": 211.9,
|
14505 |
+
"learning_rate": 9.411052631578949e-05,
|
14506 |
+
"loss": 0.0038,
|
14507 |
+
"step": 2119
|
14508 |
+
},
|
14509 |
+
{
|
14510 |
+
"epoch": 212.0,
|
14511 |
+
"learning_rate": 9.410526315789473e-05,
|
14512 |
+
"loss": 0.0056,
|
14513 |
+
"step": 2120
|
14514 |
+
},
|
14515 |
+
{
|
14516 |
+
"epoch": 212.0,
|
14517 |
+
"eval_accuracy_safe": 0.8466139297843879,
|
14518 |
+
"eval_accuracy_unlabeled": NaN,
|
14519 |
+
"eval_accuracy_unsafe": 0.996068312966864,
|
14520 |
+
"eval_iou_safe": 0.7497648583089941,
|
14521 |
+
"eval_iou_unlabeled": NaN,
|
14522 |
+
"eval_iou_unsafe": 0.9914395947822592,
|
14523 |
+
"eval_loss": 0.047027893364429474,
|
14524 |
+
"eval_mean_accuracy": 0.921341121375626,
|
14525 |
+
"eval_mean_iou": 0.8706022265456266,
|
14526 |
+
"eval_overall_accuracy": 0.9916536701259329,
|
14527 |
+
"eval_runtime": 9.6009,
|
14528 |
+
"eval_samples_per_second": 6.979,
|
14529 |
+
"eval_steps_per_second": 0.521,
|
14530 |
+
"step": 2120
|
14531 |
+
},
|
14532 |
+
{
|
14533 |
+
"epoch": 212.1,
|
14534 |
+
"learning_rate": 9.41e-05,
|
14535 |
+
"loss": 0.0045,
|
14536 |
+
"step": 2121
|
14537 |
+
},
|
14538 |
+
{
|
14539 |
+
"epoch": 212.2,
|
14540 |
+
"learning_rate": 9.409473684210527e-05,
|
14541 |
+
"loss": 0.0047,
|
14542 |
+
"step": 2122
|
14543 |
+
},
|
14544 |
+
{
|
14545 |
+
"epoch": 212.3,
|
14546 |
+
"learning_rate": 9.408947368421052e-05,
|
14547 |
+
"loss": 0.0038,
|
14548 |
+
"step": 2123
|
14549 |
+
},
|
14550 |
+
{
|
14551 |
+
"epoch": 212.4,
|
14552 |
+
"learning_rate": 9.40842105263158e-05,
|
14553 |
+
"loss": 0.0048,
|
14554 |
+
"step": 2124
|
14555 |
+
},
|
14556 |
+
{
|
14557 |
+
"epoch": 212.5,
|
14558 |
+
"learning_rate": 9.407894736842106e-05,
|
14559 |
+
"loss": 0.0055,
|
14560 |
+
"step": 2125
|
14561 |
+
},
|
14562 |
+
{
|
14563 |
+
"epoch": 212.6,
|
14564 |
+
"learning_rate": 9.407368421052632e-05,
|
14565 |
+
"loss": 0.0049,
|
14566 |
+
"step": 2126
|
14567 |
+
},
|
14568 |
+
{
|
14569 |
+
"epoch": 212.7,
|
14570 |
+
"learning_rate": 9.406842105263158e-05,
|
14571 |
+
"loss": 0.0044,
|
14572 |
+
"step": 2127
|
14573 |
+
},
|
14574 |
+
{
|
14575 |
+
"epoch": 212.8,
|
14576 |
+
"learning_rate": 9.406315789473685e-05,
|
14577 |
+
"loss": 0.0038,
|
14578 |
+
"step": 2128
|
14579 |
+
},
|
14580 |
+
{
|
14581 |
+
"epoch": 212.9,
|
14582 |
+
"learning_rate": 9.405789473684211e-05,
|
14583 |
+
"loss": 0.0046,
|
14584 |
+
"step": 2129
|
14585 |
+
},
|
14586 |
+
{
|
14587 |
+
"epoch": 213.0,
|
14588 |
+
"learning_rate": 9.405263157894737e-05,
|
14589 |
+
"loss": 0.0051,
|
14590 |
+
"step": 2130
|
14591 |
+
},
|
14592 |
+
{
|
14593 |
+
"epoch": 213.1,
|
14594 |
+
"learning_rate": 9.404736842105263e-05,
|
14595 |
+
"loss": 0.0044,
|
14596 |
+
"step": 2131
|
14597 |
+
},
|
14598 |
+
{
|
14599 |
+
"epoch": 213.2,
|
14600 |
+
"learning_rate": 9.40421052631579e-05,
|
14601 |
+
"loss": 0.0047,
|
14602 |
+
"step": 2132
|
14603 |
+
},
|
14604 |
+
{
|
14605 |
+
"epoch": 213.3,
|
14606 |
+
"learning_rate": 9.403684210526317e-05,
|
14607 |
+
"loss": 0.0049,
|
14608 |
+
"step": 2133
|
14609 |
+
},
|
14610 |
+
{
|
14611 |
+
"epoch": 213.4,
|
14612 |
+
"learning_rate": 9.403157894736842e-05,
|
14613 |
+
"loss": 0.0045,
|
14614 |
+
"step": 2134
|
14615 |
+
},
|
14616 |
+
{
|
14617 |
+
"epoch": 213.5,
|
14618 |
+
"learning_rate": 9.402631578947368e-05,
|
14619 |
+
"loss": 0.0041,
|
14620 |
+
"step": 2135
|
14621 |
+
},
|
14622 |
+
{
|
14623 |
+
"epoch": 213.6,
|
14624 |
+
"learning_rate": 9.402105263157896e-05,
|
14625 |
+
"loss": 0.0041,
|
14626 |
+
"step": 2136
|
14627 |
+
},
|
14628 |
+
{
|
14629 |
+
"epoch": 213.7,
|
14630 |
+
"learning_rate": 9.401578947368421e-05,
|
14631 |
+
"loss": 0.0057,
|
14632 |
+
"step": 2137
|
14633 |
+
},
|
14634 |
+
{
|
14635 |
+
"epoch": 213.8,
|
14636 |
+
"learning_rate": 9.401052631578948e-05,
|
14637 |
+
"loss": 0.0034,
|
14638 |
+
"step": 2138
|
14639 |
+
},
|
14640 |
+
{
|
14641 |
+
"epoch": 213.9,
|
14642 |
+
"learning_rate": 9.400526315789475e-05,
|
14643 |
+
"loss": 0.0046,
|
14644 |
+
"step": 2139
|
14645 |
+
},
|
14646 |
+
{
|
14647 |
+
"epoch": 214.0,
|
14648 |
+
"learning_rate": 9.4e-05,
|
14649 |
+
"loss": 0.0049,
|
14650 |
+
"step": 2140
|
14651 |
+
},
|
14652 |
+
{
|
14653 |
+
"epoch": 214.0,
|
14654 |
+
"eval_accuracy_safe": 0.8500294910196954,
|
14655 |
+
"eval_accuracy_unlabeled": NaN,
|
14656 |
+
"eval_accuracy_unsafe": 0.9956951796455069,
|
14657 |
+
"eval_iou_safe": 0.744704708220684,
|
14658 |
+
"eval_iou_unlabeled": NaN,
|
14659 |
+
"eval_iou_unsafe": 0.9911707597147195,
|
14660 |
+
"eval_loss": 0.048002827912569046,
|
14661 |
+
"eval_mean_accuracy": 0.9228623353326011,
|
14662 |
+
"eval_mean_iou": 0.8679377339677017,
|
14663 |
+
"eval_overall_accuracy": 0.9913924487669077,
|
14664 |
+
"eval_runtime": 9.8955,
|
14665 |
+
"eval_samples_per_second": 6.771,
|
14666 |
+
"eval_steps_per_second": 0.505,
|
14667 |
+
"step": 2140
|
14668 |
+
},
|
14669 |
+
{
|
14670 |
+
"epoch": 214.1,
|
14671 |
+
"learning_rate": 9.399473684210527e-05,
|
14672 |
+
"loss": 0.0044,
|
14673 |
+
"step": 2141
|
14674 |
+
},
|
14675 |
+
{
|
14676 |
+
"epoch": 214.2,
|
14677 |
+
"learning_rate": 9.398947368421053e-05,
|
14678 |
+
"loss": 0.0043,
|
14679 |
+
"step": 2142
|
14680 |
+
},
|
14681 |
+
{
|
14682 |
+
"epoch": 214.3,
|
14683 |
+
"learning_rate": 9.39842105263158e-05,
|
14684 |
+
"loss": 0.0043,
|
14685 |
+
"step": 2143
|
14686 |
+
},
|
14687 |
+
{
|
14688 |
+
"epoch": 214.4,
|
14689 |
+
"learning_rate": 9.397894736842106e-05,
|
14690 |
+
"loss": 0.0044,
|
14691 |
+
"step": 2144
|
14692 |
+
},
|
14693 |
+
{
|
14694 |
+
"epoch": 214.5,
|
14695 |
+
"learning_rate": 9.397368421052632e-05,
|
14696 |
+
"loss": 0.0042,
|
14697 |
+
"step": 2145
|
14698 |
+
},
|
14699 |
+
{
|
14700 |
+
"epoch": 214.6,
|
14701 |
+
"learning_rate": 9.396842105263158e-05,
|
14702 |
+
"loss": 0.0036,
|
14703 |
+
"step": 2146
|
14704 |
+
},
|
14705 |
+
{
|
14706 |
+
"epoch": 214.7,
|
14707 |
+
"learning_rate": 9.396315789473684e-05,
|
14708 |
+
"loss": 0.0048,
|
14709 |
+
"step": 2147
|
14710 |
+
},
|
14711 |
+
{
|
14712 |
+
"epoch": 214.8,
|
14713 |
+
"learning_rate": 9.39578947368421e-05,
|
14714 |
+
"loss": 0.004,
|
14715 |
+
"step": 2148
|
14716 |
+
},
|
14717 |
+
{
|
14718 |
+
"epoch": 214.9,
|
14719 |
+
"learning_rate": 9.395263157894737e-05,
|
14720 |
+
"loss": 0.0056,
|
14721 |
+
"step": 2149
|
14722 |
+
},
|
14723 |
+
{
|
14724 |
+
"epoch": 215.0,
|
14725 |
+
"learning_rate": 9.394736842105264e-05,
|
14726 |
+
"loss": 0.005,
|
14727 |
+
"step": 2150
|
14728 |
+
},
|
14729 |
+
{
|
14730 |
+
"epoch": 215.1,
|
14731 |
+
"learning_rate": 9.39421052631579e-05,
|
14732 |
+
"loss": 0.0047,
|
14733 |
+
"step": 2151
|
14734 |
+
},
|
14735 |
+
{
|
14736 |
+
"epoch": 215.2,
|
14737 |
+
"learning_rate": 9.393684210526317e-05,
|
14738 |
+
"loss": 0.0039,
|
14739 |
+
"step": 2152
|
14740 |
+
},
|
14741 |
+
{
|
14742 |
+
"epoch": 215.3,
|
14743 |
+
"learning_rate": 9.393157894736843e-05,
|
14744 |
+
"loss": 0.0043,
|
14745 |
+
"step": 2153
|
14746 |
+
},
|
14747 |
+
{
|
14748 |
+
"epoch": 215.4,
|
14749 |
+
"learning_rate": 9.392631578947368e-05,
|
14750 |
+
"loss": 0.006,
|
14751 |
+
"step": 2154
|
14752 |
+
},
|
14753 |
+
{
|
14754 |
+
"epoch": 215.5,
|
14755 |
+
"learning_rate": 9.392105263157896e-05,
|
14756 |
+
"loss": 0.0039,
|
14757 |
+
"step": 2155
|
14758 |
+
},
|
14759 |
+
{
|
14760 |
+
"epoch": 215.6,
|
14761 |
+
"learning_rate": 9.391578947368422e-05,
|
14762 |
+
"loss": 0.0054,
|
14763 |
+
"step": 2156
|
14764 |
+
},
|
14765 |
+
{
|
14766 |
+
"epoch": 215.7,
|
14767 |
+
"learning_rate": 9.391052631578948e-05,
|
14768 |
+
"loss": 0.0047,
|
14769 |
+
"step": 2157
|
14770 |
+
},
|
14771 |
+
{
|
14772 |
+
"epoch": 215.8,
|
14773 |
+
"learning_rate": 9.390526315789474e-05,
|
14774 |
+
"loss": 0.0043,
|
14775 |
+
"step": 2158
|
14776 |
+
},
|
14777 |
+
{
|
14778 |
+
"epoch": 215.9,
|
14779 |
+
"learning_rate": 9.39e-05,
|
14780 |
+
"loss": 0.005,
|
14781 |
+
"step": 2159
|
14782 |
+
},
|
14783 |
+
{
|
14784 |
+
"epoch": 216.0,
|
14785 |
+
"learning_rate": 9.389473684210527e-05,
|
14786 |
+
"loss": 0.0038,
|
14787 |
+
"step": 2160
|
14788 |
+
},
|
14789 |
+
{
|
14790 |
+
"epoch": 216.0,
|
14791 |
+
"eval_accuracy_safe": 0.8427029965189031,
|
14792 |
+
"eval_accuracy_unlabeled": NaN,
|
14793 |
+
"eval_accuracy_unsafe": 0.9961735060557309,
|
14794 |
+
"eval_iou_safe": 0.7485925235949208,
|
14795 |
+
"eval_iou_unlabeled": NaN,
|
14796 |
+
"eval_iou_unsafe": 0.9914268290945492,
|
14797 |
+
"eval_loss": 0.04742734879255295,
|
14798 |
+
"eval_mean_accuracy": 0.9194382512873169,
|
14799 |
+
"eval_mean_iou": 0.8700096763447349,
|
14800 |
+
"eval_overall_accuracy": 0.991640233281833,
|
14801 |
+
"eval_runtime": 10.4699,
|
14802 |
+
"eval_samples_per_second": 6.399,
|
14803 |
+
"eval_steps_per_second": 0.478,
|
14804 |
+
"step": 2160
|
14805 |
+
},
|
14806 |
+
{
|
14807 |
+
"epoch": 216.1,
|
14808 |
+
"learning_rate": 9.388947368421053e-05,
|
14809 |
+
"loss": 0.0056,
|
14810 |
+
"step": 2161
|
14811 |
+
},
|
14812 |
+
{
|
14813 |
+
"epoch": 216.2,
|
14814 |
+
"learning_rate": 9.388421052631579e-05,
|
14815 |
+
"loss": 0.004,
|
14816 |
+
"step": 2162
|
14817 |
+
},
|
14818 |
+
{
|
14819 |
+
"epoch": 216.3,
|
14820 |
+
"learning_rate": 9.387894736842105e-05,
|
14821 |
+
"loss": 0.0042,
|
14822 |
+
"step": 2163
|
14823 |
+
},
|
14824 |
+
{
|
14825 |
+
"epoch": 216.4,
|
14826 |
+
"learning_rate": 9.387368421052633e-05,
|
14827 |
+
"loss": 0.005,
|
14828 |
+
"step": 2164
|
14829 |
+
},
|
14830 |
+
{
|
14831 |
+
"epoch": 216.5,
|
14832 |
+
"learning_rate": 9.386842105263158e-05,
|
14833 |
+
"loss": 0.0053,
|
14834 |
+
"step": 2165
|
14835 |
+
},
|
14836 |
+
{
|
14837 |
+
"epoch": 216.6,
|
14838 |
+
"learning_rate": 9.386315789473684e-05,
|
14839 |
+
"loss": 0.0047,
|
14840 |
+
"step": 2166
|
14841 |
+
},
|
14842 |
+
{
|
14843 |
+
"epoch": 216.7,
|
14844 |
+
"learning_rate": 9.385789473684212e-05,
|
14845 |
+
"loss": 0.0046,
|
14846 |
+
"step": 2167
|
14847 |
+
},
|
14848 |
+
{
|
14849 |
+
"epoch": 216.8,
|
14850 |
+
"learning_rate": 9.385263157894737e-05,
|
14851 |
+
"loss": 0.0047,
|
14852 |
+
"step": 2168
|
14853 |
+
},
|
14854 |
+
{
|
14855 |
+
"epoch": 216.9,
|
14856 |
+
"learning_rate": 9.384736842105264e-05,
|
14857 |
+
"loss": 0.0036,
|
14858 |
+
"step": 2169
|
14859 |
+
},
|
14860 |
+
{
|
14861 |
+
"epoch": 217.0,
|
14862 |
+
"learning_rate": 9.384210526315789e-05,
|
14863 |
+
"loss": 0.004,
|
14864 |
+
"step": 2170
|
14865 |
+
},
|
14866 |
+
{
|
14867 |
+
"epoch": 217.1,
|
14868 |
+
"learning_rate": 9.383684210526317e-05,
|
14869 |
+
"loss": 0.0044,
|
14870 |
+
"step": 2171
|
14871 |
+
},
|
14872 |
+
{
|
14873 |
+
"epoch": 217.2,
|
14874 |
+
"learning_rate": 9.383157894736843e-05,
|
14875 |
+
"loss": 0.0043,
|
14876 |
+
"step": 2172
|
14877 |
+
},
|
14878 |
+
{
|
14879 |
+
"epoch": 217.3,
|
14880 |
+
"learning_rate": 9.382631578947368e-05,
|
14881 |
+
"loss": 0.0039,
|
14882 |
+
"step": 2173
|
14883 |
+
},
|
14884 |
+
{
|
14885 |
+
"epoch": 217.4,
|
14886 |
+
"learning_rate": 9.382105263157895e-05,
|
14887 |
+
"loss": 0.0048,
|
14888 |
+
"step": 2174
|
14889 |
+
},
|
14890 |
+
{
|
14891 |
+
"epoch": 217.5,
|
14892 |
+
"learning_rate": 9.381578947368422e-05,
|
14893 |
+
"loss": 0.005,
|
14894 |
+
"step": 2175
|
14895 |
+
},
|
14896 |
+
{
|
14897 |
+
"epoch": 217.6,
|
14898 |
+
"learning_rate": 9.381052631578948e-05,
|
14899 |
+
"loss": 0.0049,
|
14900 |
+
"step": 2176
|
14901 |
+
},
|
14902 |
+
{
|
14903 |
+
"epoch": 217.7,
|
14904 |
+
"learning_rate": 9.380526315789474e-05,
|
14905 |
+
"loss": 0.0047,
|
14906 |
+
"step": 2177
|
14907 |
+
},
|
14908 |
+
{
|
14909 |
+
"epoch": 217.8,
|
14910 |
+
"learning_rate": 9.38e-05,
|
14911 |
+
"loss": 0.0046,
|
14912 |
+
"step": 2178
|
14913 |
+
},
|
14914 |
+
{
|
14915 |
+
"epoch": 217.9,
|
14916 |
+
"learning_rate": 9.379473684210527e-05,
|
14917 |
+
"loss": 0.0045,
|
14918 |
+
"step": 2179
|
14919 |
+
},
|
14920 |
+
{
|
14921 |
+
"epoch": 218.0,
|
14922 |
+
"learning_rate": 9.378947368421053e-05,
|
14923 |
+
"loss": 0.0043,
|
14924 |
+
"step": 2180
|
14925 |
+
},
|
14926 |
+
{
|
14927 |
+
"epoch": 218.0,
|
14928 |
+
"eval_accuracy_safe": 0.8503456039105478,
|
14929 |
+
"eval_accuracy_unlabeled": NaN,
|
14930 |
+
"eval_accuracy_unsafe": 0.9958065329543019,
|
14931 |
+
"eval_iou_safe": 0.7473770877839123,
|
14932 |
+
"eval_iou_unlabeled": NaN,
|
14933 |
+
"eval_iou_unsafe": 0.9912911015788769,
|
14934 |
+
"eval_loss": 0.047195129096508026,
|
14935 |
+
"eval_mean_accuracy": 0.9230760684324248,
|
14936 |
+
"eval_mean_iou": 0.8693340946813946,
|
14937 |
+
"eval_overall_accuracy": 0.9915098503454408,
|
14938 |
+
"eval_runtime": 9.6744,
|
14939 |
+
"eval_samples_per_second": 6.925,
|
14940 |
+
"eval_steps_per_second": 0.517,
|
14941 |
+
"step": 2180
|
14942 |
+
},
|
14943 |
+
{
|
14944 |
+
"epoch": 218.1,
|
14945 |
+
"learning_rate": 9.378421052631579e-05,
|
14946 |
+
"loss": 0.005,
|
14947 |
+
"step": 2181
|
14948 |
+
},
|
14949 |
+
{
|
14950 |
+
"epoch": 218.2,
|
14951 |
+
"learning_rate": 9.377894736842105e-05,
|
14952 |
+
"loss": 0.0038,
|
14953 |
+
"step": 2182
|
14954 |
+
},
|
14955 |
+
{
|
14956 |
+
"epoch": 218.3,
|
14957 |
+
"learning_rate": 9.377368421052633e-05,
|
14958 |
+
"loss": 0.0048,
|
14959 |
+
"step": 2183
|
14960 |
+
},
|
14961 |
+
{
|
14962 |
+
"epoch": 218.4,
|
14963 |
+
"learning_rate": 9.376842105263158e-05,
|
14964 |
+
"loss": 0.0052,
|
14965 |
+
"step": 2184
|
14966 |
+
},
|
14967 |
+
{
|
14968 |
+
"epoch": 218.5,
|
14969 |
+
"learning_rate": 9.376315789473684e-05,
|
14970 |
+
"loss": 0.0053,
|
14971 |
+
"step": 2185
|
14972 |
+
},
|
14973 |
+
{
|
14974 |
+
"epoch": 218.6,
|
14975 |
+
"learning_rate": 9.375789473684212e-05,
|
14976 |
+
"loss": 0.0035,
|
14977 |
+
"step": 2186
|
14978 |
+
},
|
14979 |
+
{
|
14980 |
+
"epoch": 218.7,
|
14981 |
+
"learning_rate": 9.375263157894736e-05,
|
14982 |
+
"loss": 0.005,
|
14983 |
+
"step": 2187
|
14984 |
+
},
|
14985 |
+
{
|
14986 |
+
"epoch": 218.8,
|
14987 |
+
"learning_rate": 9.374736842105264e-05,
|
14988 |
+
"loss": 0.0047,
|
14989 |
+
"step": 2188
|
14990 |
+
},
|
14991 |
+
{
|
14992 |
+
"epoch": 218.9,
|
14993 |
+
"learning_rate": 9.37421052631579e-05,
|
14994 |
+
"loss": 0.0042,
|
14995 |
+
"step": 2189
|
14996 |
+
},
|
14997 |
+
{
|
14998 |
+
"epoch": 219.0,
|
14999 |
+
"learning_rate": 9.373684210526316e-05,
|
15000 |
+
"loss": 0.0046,
|
15001 |
+
"step": 2190
|
15002 |
+
},
|
15003 |
+
{
|
15004 |
+
"epoch": 219.1,
|
15005 |
+
"learning_rate": 9.373157894736843e-05,
|
15006 |
+
"loss": 0.0045,
|
15007 |
+
"step": 2191
|
15008 |
+
},
|
15009 |
+
{
|
15010 |
+
"epoch": 219.2,
|
15011 |
+
"learning_rate": 9.372631578947369e-05,
|
15012 |
+
"loss": 0.0056,
|
15013 |
+
"step": 2192
|
15014 |
+
},
|
15015 |
+
{
|
15016 |
+
"epoch": 219.3,
|
15017 |
+
"learning_rate": 9.372105263157895e-05,
|
15018 |
+
"loss": 0.0043,
|
15019 |
+
"step": 2193
|
15020 |
+
},
|
15021 |
+
{
|
15022 |
+
"epoch": 219.4,
|
15023 |
+
"learning_rate": 9.371578947368421e-05,
|
15024 |
+
"loss": 0.0035,
|
15025 |
+
"step": 2194
|
15026 |
+
},
|
15027 |
+
{
|
15028 |
+
"epoch": 219.5,
|
15029 |
+
"learning_rate": 9.371052631578948e-05,
|
15030 |
+
"loss": 0.0052,
|
15031 |
+
"step": 2195
|
15032 |
+
},
|
15033 |
+
{
|
15034 |
+
"epoch": 219.6,
|
15035 |
+
"learning_rate": 9.370526315789474e-05,
|
15036 |
+
"loss": 0.0045,
|
15037 |
+
"step": 2196
|
15038 |
+
},
|
15039 |
+
{
|
15040 |
+
"epoch": 219.7,
|
15041 |
+
"learning_rate": 9.370000000000001e-05,
|
15042 |
+
"loss": 0.005,
|
15043 |
+
"step": 2197
|
15044 |
+
},
|
15045 |
+
{
|
15046 |
+
"epoch": 219.8,
|
15047 |
+
"learning_rate": 9.369473684210526e-05,
|
15048 |
+
"loss": 0.0047,
|
15049 |
+
"step": 2198
|
15050 |
+
},
|
15051 |
+
{
|
15052 |
+
"epoch": 219.9,
|
15053 |
+
"learning_rate": 9.368947368421053e-05,
|
15054 |
+
"loss": 0.004,
|
15055 |
+
"step": 2199
|
15056 |
+
},
|
15057 |
+
{
|
15058 |
+
"epoch": 220.0,
|
15059 |
+
"learning_rate": 9.36842105263158e-05,
|
15060 |
+
"loss": 0.005,
|
15061 |
+
"step": 2200
|
15062 |
+
},
|
15063 |
+
{
|
15064 |
+
"epoch": 220.0,
|
15065 |
+
"eval_accuracy_safe": 0.8347596963774234,
|
15066 |
+
"eval_accuracy_unlabeled": NaN,
|
15067 |
+
"eval_accuracy_unsafe": 0.9964474891706268,
|
15068 |
+
"eval_iou_safe": 0.7475136099862951,
|
15069 |
+
"eval_iou_unlabeled": NaN,
|
15070 |
+
"eval_iou_unsafe": 0.991460939456024,
|
15071 |
+
"eval_loss": 0.04711836576461792,
|
15072 |
+
"eval_mean_accuracy": 0.9156035927740251,
|
15073 |
+
"eval_mean_iou": 0.8694872747211595,
|
15074 |
+
"eval_overall_accuracy": 0.9916714910250991,
|
15075 |
+
"eval_runtime": 9.7724,
|
15076 |
+
"eval_samples_per_second": 6.856,
|
15077 |
+
"eval_steps_per_second": 0.512,
|
15078 |
+
"step": 2200
|
15079 |
+
},
|
15080 |
+
{
|
15081 |
+
"epoch": 220.1,
|
15082 |
+
"learning_rate": 9.367894736842105e-05,
|
15083 |
+
"loss": 0.0049,
|
15084 |
+
"step": 2201
|
15085 |
+
},
|
15086 |
+
{
|
15087 |
+
"epoch": 220.2,
|
15088 |
+
"learning_rate": 9.367368421052633e-05,
|
15089 |
+
"loss": 0.0038,
|
15090 |
+
"step": 2202
|
15091 |
+
},
|
15092 |
+
{
|
15093 |
+
"epoch": 220.3,
|
15094 |
+
"learning_rate": 9.366842105263159e-05,
|
15095 |
+
"loss": 0.0041,
|
15096 |
+
"step": 2203
|
15097 |
+
},
|
15098 |
+
{
|
15099 |
+
"epoch": 220.4,
|
15100 |
+
"learning_rate": 9.366315789473684e-05,
|
15101 |
+
"loss": 0.0043,
|
15102 |
+
"step": 2204
|
15103 |
+
},
|
15104 |
+
{
|
15105 |
+
"epoch": 220.5,
|
15106 |
+
"learning_rate": 9.365789473684211e-05,
|
15107 |
+
"loss": 0.0045,
|
15108 |
+
"step": 2205
|
15109 |
+
},
|
15110 |
+
{
|
15111 |
+
"epoch": 220.6,
|
15112 |
+
"learning_rate": 9.365263157894738e-05,
|
15113 |
+
"loss": 0.0041,
|
15114 |
+
"step": 2206
|
15115 |
+
},
|
15116 |
+
{
|
15117 |
+
"epoch": 220.7,
|
15118 |
+
"learning_rate": 9.364736842105264e-05,
|
15119 |
+
"loss": 0.0047,
|
15120 |
+
"step": 2207
|
15121 |
+
},
|
15122 |
+
{
|
15123 |
+
"epoch": 220.8,
|
15124 |
+
"learning_rate": 9.36421052631579e-05,
|
15125 |
+
"loss": 0.0046,
|
15126 |
+
"step": 2208
|
15127 |
+
},
|
15128 |
+
{
|
15129 |
+
"epoch": 220.9,
|
15130 |
+
"learning_rate": 9.363684210526316e-05,
|
15131 |
+
"loss": 0.0047,
|
15132 |
+
"step": 2209
|
15133 |
+
},
|
15134 |
+
{
|
15135 |
+
"epoch": 221.0,
|
15136 |
+
"learning_rate": 9.363157894736842e-05,
|
15137 |
+
"loss": 0.0045,
|
15138 |
+
"step": 2210
|
15139 |
+
},
|
15140 |
+
{
|
15141 |
+
"epoch": 221.1,
|
15142 |
+
"learning_rate": 9.362631578947369e-05,
|
15143 |
+
"loss": 0.0043,
|
15144 |
+
"step": 2211
|
15145 |
+
},
|
15146 |
+
{
|
15147 |
+
"epoch": 221.2,
|
15148 |
+
"learning_rate": 9.362105263157895e-05,
|
15149 |
+
"loss": 0.004,
|
15150 |
+
"step": 2212
|
15151 |
+
},
|
15152 |
+
{
|
15153 |
+
"epoch": 221.3,
|
15154 |
+
"learning_rate": 9.361578947368421e-05,
|
15155 |
+
"loss": 0.0044,
|
15156 |
+
"step": 2213
|
15157 |
+
},
|
15158 |
+
{
|
15159 |
+
"epoch": 221.4,
|
15160 |
+
"learning_rate": 9.361052631578949e-05,
|
15161 |
+
"loss": 0.0047,
|
15162 |
+
"step": 2214
|
15163 |
+
},
|
15164 |
+
{
|
15165 |
+
"epoch": 221.5,
|
15166 |
+
"learning_rate": 9.360526315789474e-05,
|
15167 |
+
"loss": 0.0045,
|
15168 |
+
"step": 2215
|
15169 |
+
},
|
15170 |
+
{
|
15171 |
+
"epoch": 221.6,
|
15172 |
+
"learning_rate": 9.360000000000001e-05,
|
15173 |
+
"loss": 0.0043,
|
15174 |
+
"step": 2216
|
15175 |
+
},
|
15176 |
+
{
|
15177 |
+
"epoch": 221.7,
|
15178 |
+
"learning_rate": 9.359473684210527e-05,
|
15179 |
+
"loss": 0.0054,
|
15180 |
+
"step": 2217
|
15181 |
+
},
|
15182 |
+
{
|
15183 |
+
"epoch": 221.8,
|
15184 |
+
"learning_rate": 9.358947368421052e-05,
|
15185 |
+
"loss": 0.0043,
|
15186 |
+
"step": 2218
|
15187 |
+
},
|
15188 |
+
{
|
15189 |
+
"epoch": 221.9,
|
15190 |
+
"learning_rate": 9.35842105263158e-05,
|
15191 |
+
"loss": 0.004,
|
15192 |
+
"step": 2219
|
15193 |
+
},
|
15194 |
+
{
|
15195 |
+
"epoch": 222.0,
|
15196 |
+
"learning_rate": 9.357894736842106e-05,
|
15197 |
+
"loss": 0.0041,
|
15198 |
+
"step": 2220
|
15199 |
+
},
|
15200 |
+
{
|
15201 |
+
"epoch": 222.0,
|
15202 |
+
"eval_accuracy_safe": 0.841080026676844,
|
15203 |
+
"eval_accuracy_unlabeled": NaN,
|
15204 |
+
"eval_accuracy_unsafe": 0.9964024902307712,
|
15205 |
+
"eval_iou_safe": 0.7521775629566935,
|
15206 |
+
"eval_iou_unlabeled": NaN,
|
15207 |
+
"eval_iou_unsafe": 0.991605971147573,
|
15208 |
+
"eval_loss": 0.047169264405965805,
|
15209 |
+
"eval_mean_accuracy": 0.9187412584538076,
|
15210 |
+
"eval_mean_iou": 0.8718917670521332,
|
15211 |
+
"eval_overall_accuracy": 0.99181451370467,
|
15212 |
+
"eval_runtime": 10.9187,
|
15213 |
+
"eval_samples_per_second": 6.136,
|
15214 |
+
"eval_steps_per_second": 0.458,
|
15215 |
+
"step": 2220
|
15216 |
+
},
|
15217 |
+
{
|
15218 |
+
"epoch": 222.1,
|
15219 |
+
"learning_rate": 9.357368421052632e-05,
|
15220 |
+
"loss": 0.004,
|
15221 |
+
"step": 2221
|
15222 |
+
},
|
15223 |
+
{
|
15224 |
+
"epoch": 222.2,
|
15225 |
+
"learning_rate": 9.356842105263159e-05,
|
15226 |
+
"loss": 0.0041,
|
15227 |
+
"step": 2222
|
15228 |
+
},
|
15229 |
+
{
|
15230 |
+
"epoch": 222.3,
|
15231 |
+
"learning_rate": 9.356315789473685e-05,
|
15232 |
+
"loss": 0.0042,
|
15233 |
+
"step": 2223
|
15234 |
+
},
|
15235 |
+
{
|
15236 |
+
"epoch": 222.4,
|
15237 |
+
"learning_rate": 9.355789473684211e-05,
|
15238 |
+
"loss": 0.005,
|
15239 |
+
"step": 2224
|
15240 |
+
},
|
15241 |
+
{
|
15242 |
+
"epoch": 222.5,
|
15243 |
+
"learning_rate": 9.355263157894737e-05,
|
15244 |
+
"loss": 0.0039,
|
15245 |
+
"step": 2225
|
15246 |
+
},
|
15247 |
+
{
|
15248 |
+
"epoch": 222.6,
|
15249 |
+
"learning_rate": 9.354736842105264e-05,
|
15250 |
+
"loss": 0.0052,
|
15251 |
+
"step": 2226
|
15252 |
+
},
|
15253 |
+
{
|
15254 |
+
"epoch": 222.7,
|
15255 |
+
"learning_rate": 9.35421052631579e-05,
|
15256 |
+
"loss": 0.0043,
|
15257 |
+
"step": 2227
|
15258 |
+
},
|
15259 |
+
{
|
15260 |
+
"epoch": 222.8,
|
15261 |
+
"learning_rate": 9.353684210526317e-05,
|
15262 |
+
"loss": 0.004,
|
15263 |
+
"step": 2228
|
15264 |
+
},
|
15265 |
+
{
|
15266 |
+
"epoch": 222.9,
|
15267 |
+
"learning_rate": 9.353157894736842e-05,
|
15268 |
+
"loss": 0.0043,
|
15269 |
+
"step": 2229
|
15270 |
+
},
|
15271 |
+
{
|
15272 |
+
"epoch": 223.0,
|
15273 |
+
"learning_rate": 9.352631578947368e-05,
|
15274 |
+
"loss": 0.0046,
|
15275 |
+
"step": 2230
|
15276 |
+
},
|
15277 |
+
{
|
15278 |
+
"epoch": 223.1,
|
15279 |
+
"learning_rate": 9.352105263157896e-05,
|
15280 |
+
"loss": 0.0037,
|
15281 |
+
"step": 2231
|
15282 |
+
},
|
15283 |
+
{
|
15284 |
+
"epoch": 223.2,
|
15285 |
+
"learning_rate": 9.351578947368421e-05,
|
15286 |
+
"loss": 0.0038,
|
15287 |
+
"step": 2232
|
15288 |
+
},
|
15289 |
+
{
|
15290 |
+
"epoch": 223.3,
|
15291 |
+
"learning_rate": 9.351052631578948e-05,
|
15292 |
+
"loss": 0.0045,
|
15293 |
+
"step": 2233
|
15294 |
+
},
|
15295 |
+
{
|
15296 |
+
"epoch": 223.4,
|
15297 |
+
"learning_rate": 9.350526315789473e-05,
|
15298 |
+
"loss": 0.0039,
|
15299 |
+
"step": 2234
|
15300 |
+
},
|
15301 |
+
{
|
15302 |
+
"epoch": 223.5,
|
15303 |
+
"learning_rate": 9.350000000000001e-05,
|
15304 |
+
"loss": 0.0056,
|
15305 |
+
"step": 2235
|
15306 |
+
},
|
15307 |
+
{
|
15308 |
+
"epoch": 223.6,
|
15309 |
+
"learning_rate": 9.349473684210527e-05,
|
15310 |
+
"loss": 0.0046,
|
15311 |
+
"step": 2236
|
15312 |
+
},
|
15313 |
+
{
|
15314 |
+
"epoch": 223.7,
|
15315 |
+
"learning_rate": 9.348947368421052e-05,
|
15316 |
+
"loss": 0.0043,
|
15317 |
+
"step": 2237
|
15318 |
+
},
|
15319 |
+
{
|
15320 |
+
"epoch": 223.8,
|
15321 |
+
"learning_rate": 9.34842105263158e-05,
|
15322 |
+
"loss": 0.0042,
|
15323 |
+
"step": 2238
|
15324 |
+
},
|
15325 |
+
{
|
15326 |
+
"epoch": 223.9,
|
15327 |
+
"learning_rate": 9.347894736842106e-05,
|
15328 |
+
"loss": 0.0044,
|
15329 |
+
"step": 2239
|
15330 |
+
},
|
15331 |
+
{
|
15332 |
+
"epoch": 224.0,
|
15333 |
+
"learning_rate": 9.347368421052632e-05,
|
15334 |
+
"loss": 0.0041,
|
15335 |
+
"step": 2240
|
15336 |
+
},
|
15337 |
+
{
|
15338 |
+
"epoch": 224.0,
|
15339 |
+
"eval_accuracy_safe": 0.8476605718559296,
|
15340 |
+
"eval_accuracy_unlabeled": NaN,
|
15341 |
+
"eval_accuracy_unsafe": 0.9961528546517815,
|
15342 |
+
"eval_iou_safe": 0.7525428833490766,
|
15343 |
+
"eval_iou_unlabeled": NaN,
|
15344 |
+
"eval_iou_unsafe": 0.9915551849230432,
|
15345 |
+
"eval_loss": 0.04713202640414238,
|
15346 |
+
"eval_mean_accuracy": 0.9219067132538555,
|
15347 |
+
"eval_mean_iou": 0.8720490341360598,
|
15348 |
+
"eval_overall_accuracy": 0.9917666307136194,
|
15349 |
+
"eval_runtime": 10.3365,
|
15350 |
+
"eval_samples_per_second": 6.482,
|
15351 |
+
"eval_steps_per_second": 0.484,
|
15352 |
+
"step": 2240
|
15353 |
}
|
15354 |
],
|
15355 |
"max_steps": 20000,
|
15356 |
"num_train_epochs": 2000,
|
15357 |
+
"total_flos": 1.376984980104885e+19,
|
15358 |
"trial_name": null,
|
15359 |
"trial_params": null
|
15360 |
}
|
{checkpoint-1920 β checkpoint-2240}/training_args.bin
RENAMED
File without changes
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5488ff60e968739c54eef6c3eddadc290534ca260635011c03a8b33733366b1d
|
3 |
size 338796489
|
runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c05e5e72964c837469b261cd2e53d4b82400f6580dd19e7d28b9faa78c85f67a
|
3 |
+
size 443497
|