Training in progress, step 2080
Browse files- {checkpoint-1760 β checkpoint-2080}/config.json +0 -0
- {checkpoint-1760 β checkpoint-2080}/optimizer.pt +1 -1
- {checkpoint-1760 β checkpoint-2080}/pytorch_model.bin +1 -1
- {checkpoint-1760 β checkpoint-2080}/rng_state.pth +1 -1
- {checkpoint-1760 β checkpoint-2080}/scheduler.pt +1 -1
- {checkpoint-1760 β checkpoint-2080}/trainer_state.json +2195 -3
- {checkpoint-1760 β checkpoint-2080}/training_args.bin +0 -0
- pytorch_model.bin +1 -1
- runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0 +2 -2
{checkpoint-1760 β checkpoint-2080}/config.json
RENAMED
File without changes
|
{checkpoint-1760 β checkpoint-2080}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 677456319
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e1ebd96b3b27c14e5d7471c0ec9d23ef1802cc1c5b550b545a50563c6a6d338
|
3 |
size 677456319
|
{checkpoint-1760 β checkpoint-2080}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:601825867ffe486a1fe2c2d6445b63bf2334be4c983b99fd941f121a72f4c661
|
3 |
size 338796489
|
{checkpoint-1760 β checkpoint-2080}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:23ca27a1a06e402df89e614592c5e35dfe8b4b1134a4ce8376808aa60c5f1963
|
3 |
size 14575
|
{checkpoint-1760 β checkpoint-2080}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba71d8953da2654bf003cab6573037cc1edd5cdfa4fe5b673fcf5e5d650c7b91
|
3 |
size 627
|
{checkpoint-1760 β checkpoint-2080}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -12062,11 +12062,2203 @@
|
|
12062 |
"eval_samples_per_second": 6.987,
|
12063 |
"eval_steps_per_second": 0.521,
|
12064 |
"step": 1760
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12065 |
}
|
12066 |
],
|
12067 |
"max_steps": 20000,
|
12068 |
"num_train_epochs": 2000,
|
12069 |
-
"total_flos": 1.
|
12070 |
"trial_name": null,
|
12071 |
"trial_params": null
|
12072 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
+
"epoch": 208.0,
|
5 |
+
"global_step": 2080,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
12062 |
"eval_samples_per_second": 6.987,
|
12063 |
"eval_steps_per_second": 0.521,
|
12064 |
"step": 1760
|
12065 |
+
},
|
12066 |
+
{
|
12067 |
+
"epoch": 176.1,
|
12068 |
+
"learning_rate": 9.599473684210526e-05,
|
12069 |
+
"loss": 0.0055,
|
12070 |
+
"step": 1761
|
12071 |
+
},
|
12072 |
+
{
|
12073 |
+
"epoch": 176.2,
|
12074 |
+
"learning_rate": 9.598947368421053e-05,
|
12075 |
+
"loss": 0.0055,
|
12076 |
+
"step": 1762
|
12077 |
+
},
|
12078 |
+
{
|
12079 |
+
"epoch": 176.3,
|
12080 |
+
"learning_rate": 9.59842105263158e-05,
|
12081 |
+
"loss": 0.0068,
|
12082 |
+
"step": 1763
|
12083 |
+
},
|
12084 |
+
{
|
12085 |
+
"epoch": 176.4,
|
12086 |
+
"learning_rate": 9.597894736842105e-05,
|
12087 |
+
"loss": 0.0062,
|
12088 |
+
"step": 1764
|
12089 |
+
},
|
12090 |
+
{
|
12091 |
+
"epoch": 176.5,
|
12092 |
+
"learning_rate": 9.597368421052631e-05,
|
12093 |
+
"loss": 0.0062,
|
12094 |
+
"step": 1765
|
12095 |
+
},
|
12096 |
+
{
|
12097 |
+
"epoch": 176.6,
|
12098 |
+
"learning_rate": 9.596842105263159e-05,
|
12099 |
+
"loss": 0.005,
|
12100 |
+
"step": 1766
|
12101 |
+
},
|
12102 |
+
{
|
12103 |
+
"epoch": 176.7,
|
12104 |
+
"learning_rate": 9.596315789473684e-05,
|
12105 |
+
"loss": 0.0058,
|
12106 |
+
"step": 1767
|
12107 |
+
},
|
12108 |
+
{
|
12109 |
+
"epoch": 176.8,
|
12110 |
+
"learning_rate": 9.595789473684211e-05,
|
12111 |
+
"loss": 0.0048,
|
12112 |
+
"step": 1768
|
12113 |
+
},
|
12114 |
+
{
|
12115 |
+
"epoch": 176.9,
|
12116 |
+
"learning_rate": 9.595263157894738e-05,
|
12117 |
+
"loss": 0.0053,
|
12118 |
+
"step": 1769
|
12119 |
+
},
|
12120 |
+
{
|
12121 |
+
"epoch": 177.0,
|
12122 |
+
"learning_rate": 9.594736842105264e-05,
|
12123 |
+
"loss": 0.005,
|
12124 |
+
"step": 1770
|
12125 |
+
},
|
12126 |
+
{
|
12127 |
+
"epoch": 177.1,
|
12128 |
+
"learning_rate": 9.59421052631579e-05,
|
12129 |
+
"loss": 0.0059,
|
12130 |
+
"step": 1771
|
12131 |
+
},
|
12132 |
+
{
|
12133 |
+
"epoch": 177.2,
|
12134 |
+
"learning_rate": 9.593684210526316e-05,
|
12135 |
+
"loss": 0.0049,
|
12136 |
+
"step": 1772
|
12137 |
+
},
|
12138 |
+
{
|
12139 |
+
"epoch": 177.3,
|
12140 |
+
"learning_rate": 9.593157894736843e-05,
|
12141 |
+
"loss": 0.0055,
|
12142 |
+
"step": 1773
|
12143 |
+
},
|
12144 |
+
{
|
12145 |
+
"epoch": 177.4,
|
12146 |
+
"learning_rate": 9.592631578947369e-05,
|
12147 |
+
"loss": 0.007,
|
12148 |
+
"step": 1774
|
12149 |
+
},
|
12150 |
+
{
|
12151 |
+
"epoch": 177.5,
|
12152 |
+
"learning_rate": 9.592105263157895e-05,
|
12153 |
+
"loss": 0.0062,
|
12154 |
+
"step": 1775
|
12155 |
+
},
|
12156 |
+
{
|
12157 |
+
"epoch": 177.6,
|
12158 |
+
"learning_rate": 9.591578947368421e-05,
|
12159 |
+
"loss": 0.0051,
|
12160 |
+
"step": 1776
|
12161 |
+
},
|
12162 |
+
{
|
12163 |
+
"epoch": 177.7,
|
12164 |
+
"learning_rate": 9.591052631578948e-05,
|
12165 |
+
"loss": 0.0053,
|
12166 |
+
"step": 1777
|
12167 |
+
},
|
12168 |
+
{
|
12169 |
+
"epoch": 177.8,
|
12170 |
+
"learning_rate": 9.590526315789474e-05,
|
12171 |
+
"loss": 0.0049,
|
12172 |
+
"step": 1778
|
12173 |
+
},
|
12174 |
+
{
|
12175 |
+
"epoch": 177.9,
|
12176 |
+
"learning_rate": 9.59e-05,
|
12177 |
+
"loss": 0.0051,
|
12178 |
+
"step": 1779
|
12179 |
+
},
|
12180 |
+
{
|
12181 |
+
"epoch": 178.0,
|
12182 |
+
"learning_rate": 9.589473684210528e-05,
|
12183 |
+
"loss": 0.0061,
|
12184 |
+
"step": 1780
|
12185 |
+
},
|
12186 |
+
{
|
12187 |
+
"epoch": 178.0,
|
12188 |
+
"eval_accuracy_safe": 0.8491833107813771,
|
12189 |
+
"eval_accuracy_unlabeled": NaN,
|
12190 |
+
"eval_accuracy_unsafe": 0.9960346957666851,
|
12191 |
+
"eval_iou_safe": 0.7513054451815345,
|
12192 |
+
"eval_iou_unlabeled": NaN,
|
12193 |
+
"eval_iou_unsafe": 0.9914833128833639,
|
12194 |
+
"eval_loss": 0.04173571988940239,
|
12195 |
+
"eval_mean_accuracy": 0.9226090032740311,
|
12196 |
+
"eval_mean_iou": 0.8713943790324492,
|
12197 |
+
"eval_overall_accuracy": 0.9916969413187966,
|
12198 |
+
"eval_runtime": 9.6891,
|
12199 |
+
"eval_samples_per_second": 6.915,
|
12200 |
+
"eval_steps_per_second": 0.516,
|
12201 |
+
"step": 1780
|
12202 |
+
},
|
12203 |
+
{
|
12204 |
+
"epoch": 178.1,
|
12205 |
+
"learning_rate": 9.588947368421052e-05,
|
12206 |
+
"loss": 0.0059,
|
12207 |
+
"step": 1781
|
12208 |
+
},
|
12209 |
+
{
|
12210 |
+
"epoch": 178.2,
|
12211 |
+
"learning_rate": 9.58842105263158e-05,
|
12212 |
+
"loss": 0.006,
|
12213 |
+
"step": 1782
|
12214 |
+
},
|
12215 |
+
{
|
12216 |
+
"epoch": 178.3,
|
12217 |
+
"learning_rate": 9.587894736842106e-05,
|
12218 |
+
"loss": 0.0043,
|
12219 |
+
"step": 1783
|
12220 |
+
},
|
12221 |
+
{
|
12222 |
+
"epoch": 178.4,
|
12223 |
+
"learning_rate": 9.587368421052631e-05,
|
12224 |
+
"loss": 0.0054,
|
12225 |
+
"step": 1784
|
12226 |
+
},
|
12227 |
+
{
|
12228 |
+
"epoch": 178.5,
|
12229 |
+
"learning_rate": 9.586842105263159e-05,
|
12230 |
+
"loss": 0.0069,
|
12231 |
+
"step": 1785
|
12232 |
+
},
|
12233 |
+
{
|
12234 |
+
"epoch": 178.6,
|
12235 |
+
"learning_rate": 9.586315789473685e-05,
|
12236 |
+
"loss": 0.0051,
|
12237 |
+
"step": 1786
|
12238 |
+
},
|
12239 |
+
{
|
12240 |
+
"epoch": 178.7,
|
12241 |
+
"learning_rate": 9.585789473684211e-05,
|
12242 |
+
"loss": 0.0048,
|
12243 |
+
"step": 1787
|
12244 |
+
},
|
12245 |
+
{
|
12246 |
+
"epoch": 178.8,
|
12247 |
+
"learning_rate": 9.585263157894737e-05,
|
12248 |
+
"loss": 0.0054,
|
12249 |
+
"step": 1788
|
12250 |
+
},
|
12251 |
+
{
|
12252 |
+
"epoch": 178.9,
|
12253 |
+
"learning_rate": 9.584736842105264e-05,
|
12254 |
+
"loss": 0.0058,
|
12255 |
+
"step": 1789
|
12256 |
+
},
|
12257 |
+
{
|
12258 |
+
"epoch": 179.0,
|
12259 |
+
"learning_rate": 9.58421052631579e-05,
|
12260 |
+
"loss": 0.0053,
|
12261 |
+
"step": 1790
|
12262 |
+
},
|
12263 |
+
{
|
12264 |
+
"epoch": 179.1,
|
12265 |
+
"learning_rate": 9.583684210526316e-05,
|
12266 |
+
"loss": 0.0069,
|
12267 |
+
"step": 1791
|
12268 |
+
},
|
12269 |
+
{
|
12270 |
+
"epoch": 179.2,
|
12271 |
+
"learning_rate": 9.583157894736842e-05,
|
12272 |
+
"loss": 0.0052,
|
12273 |
+
"step": 1792
|
12274 |
+
},
|
12275 |
+
{
|
12276 |
+
"epoch": 179.3,
|
12277 |
+
"learning_rate": 9.582631578947369e-05,
|
12278 |
+
"loss": 0.0042,
|
12279 |
+
"step": 1793
|
12280 |
+
},
|
12281 |
+
{
|
12282 |
+
"epoch": 179.4,
|
12283 |
+
"learning_rate": 9.582105263157896e-05,
|
12284 |
+
"loss": 0.0053,
|
12285 |
+
"step": 1794
|
12286 |
+
},
|
12287 |
+
{
|
12288 |
+
"epoch": 179.5,
|
12289 |
+
"learning_rate": 9.581578947368421e-05,
|
12290 |
+
"loss": 0.0057,
|
12291 |
+
"step": 1795
|
12292 |
+
},
|
12293 |
+
{
|
12294 |
+
"epoch": 179.6,
|
12295 |
+
"learning_rate": 9.581052631578947e-05,
|
12296 |
+
"loss": 0.0044,
|
12297 |
+
"step": 1796
|
12298 |
+
},
|
12299 |
+
{
|
12300 |
+
"epoch": 179.7,
|
12301 |
+
"learning_rate": 9.580526315789475e-05,
|
12302 |
+
"loss": 0.0061,
|
12303 |
+
"step": 1797
|
12304 |
+
},
|
12305 |
+
{
|
12306 |
+
"epoch": 179.8,
|
12307 |
+
"learning_rate": 9.58e-05,
|
12308 |
+
"loss": 0.0059,
|
12309 |
+
"step": 1798
|
12310 |
+
},
|
12311 |
+
{
|
12312 |
+
"epoch": 179.9,
|
12313 |
+
"learning_rate": 9.579473684210527e-05,
|
12314 |
+
"loss": 0.0048,
|
12315 |
+
"step": 1799
|
12316 |
+
},
|
12317 |
+
{
|
12318 |
+
"epoch": 180.0,
|
12319 |
+
"learning_rate": 9.578947368421052e-05,
|
12320 |
+
"loss": 0.0061,
|
12321 |
+
"step": 1800
|
12322 |
+
},
|
12323 |
+
{
|
12324 |
+
"epoch": 180.0,
|
12325 |
+
"eval_accuracy_safe": 0.8537785128044996,
|
12326 |
+
"eval_accuracy_unlabeled": NaN,
|
12327 |
+
"eval_accuracy_unsafe": 0.9959957983779965,
|
12328 |
+
"eval_iou_safe": 0.7545179055383415,
|
12329 |
+
"eval_iou_unlabeled": NaN,
|
12330 |
+
"eval_iou_unsafe": 0.9915826485192842,
|
12331 |
+
"eval_loss": 0.04071439057588577,
|
12332 |
+
"eval_mean_accuracy": 0.924887155591248,
|
12333 |
+
"eval_mean_iou": 0.8730502770288129,
|
12334 |
+
"eval_overall_accuracy": 0.9917949277963212,
|
12335 |
+
"eval_runtime": 9.6123,
|
12336 |
+
"eval_samples_per_second": 6.97,
|
12337 |
+
"eval_steps_per_second": 0.52,
|
12338 |
+
"step": 1800
|
12339 |
+
},
|
12340 |
+
{
|
12341 |
+
"epoch": 180.1,
|
12342 |
+
"learning_rate": 9.57842105263158e-05,
|
12343 |
+
"loss": 0.0048,
|
12344 |
+
"step": 1801
|
12345 |
+
},
|
12346 |
+
{
|
12347 |
+
"epoch": 180.2,
|
12348 |
+
"learning_rate": 9.577894736842106e-05,
|
12349 |
+
"loss": 0.0048,
|
12350 |
+
"step": 1802
|
12351 |
+
},
|
12352 |
+
{
|
12353 |
+
"epoch": 180.3,
|
12354 |
+
"learning_rate": 9.577368421052631e-05,
|
12355 |
+
"loss": 0.0049,
|
12356 |
+
"step": 1803
|
12357 |
+
},
|
12358 |
+
{
|
12359 |
+
"epoch": 180.4,
|
12360 |
+
"learning_rate": 9.576842105263159e-05,
|
12361 |
+
"loss": 0.0055,
|
12362 |
+
"step": 1804
|
12363 |
+
},
|
12364 |
+
{
|
12365 |
+
"epoch": 180.5,
|
12366 |
+
"learning_rate": 9.576315789473685e-05,
|
12367 |
+
"loss": 0.0056,
|
12368 |
+
"step": 1805
|
12369 |
+
},
|
12370 |
+
{
|
12371 |
+
"epoch": 180.6,
|
12372 |
+
"learning_rate": 9.575789473684211e-05,
|
12373 |
+
"loss": 0.0052,
|
12374 |
+
"step": 1806
|
12375 |
+
},
|
12376 |
+
{
|
12377 |
+
"epoch": 180.7,
|
12378 |
+
"learning_rate": 9.575263157894737e-05,
|
12379 |
+
"loss": 0.005,
|
12380 |
+
"step": 1807
|
12381 |
+
},
|
12382 |
+
{
|
12383 |
+
"epoch": 180.8,
|
12384 |
+
"learning_rate": 9.574736842105265e-05,
|
12385 |
+
"loss": 0.005,
|
12386 |
+
"step": 1808
|
12387 |
+
},
|
12388 |
+
{
|
12389 |
+
"epoch": 180.9,
|
12390 |
+
"learning_rate": 9.57421052631579e-05,
|
12391 |
+
"loss": 0.0072,
|
12392 |
+
"step": 1809
|
12393 |
+
},
|
12394 |
+
{
|
12395 |
+
"epoch": 181.0,
|
12396 |
+
"learning_rate": 9.573684210526316e-05,
|
12397 |
+
"loss": 0.0057,
|
12398 |
+
"step": 1810
|
12399 |
+
},
|
12400 |
+
{
|
12401 |
+
"epoch": 181.1,
|
12402 |
+
"learning_rate": 9.573157894736842e-05,
|
12403 |
+
"loss": 0.0046,
|
12404 |
+
"step": 1811
|
12405 |
+
},
|
12406 |
+
{
|
12407 |
+
"epoch": 181.2,
|
12408 |
+
"learning_rate": 9.572631578947368e-05,
|
12409 |
+
"loss": 0.0059,
|
12410 |
+
"step": 1812
|
12411 |
+
},
|
12412 |
+
{
|
12413 |
+
"epoch": 181.3,
|
12414 |
+
"learning_rate": 9.572105263157896e-05,
|
12415 |
+
"loss": 0.006,
|
12416 |
+
"step": 1813
|
12417 |
+
},
|
12418 |
+
{
|
12419 |
+
"epoch": 181.4,
|
12420 |
+
"learning_rate": 9.571578947368421e-05,
|
12421 |
+
"loss": 0.0047,
|
12422 |
+
"step": 1814
|
12423 |
+
},
|
12424 |
+
{
|
12425 |
+
"epoch": 181.5,
|
12426 |
+
"learning_rate": 9.571052631578948e-05,
|
12427 |
+
"loss": 0.0053,
|
12428 |
+
"step": 1815
|
12429 |
+
},
|
12430 |
+
{
|
12431 |
+
"epoch": 181.6,
|
12432 |
+
"learning_rate": 9.570526315789475e-05,
|
12433 |
+
"loss": 0.0065,
|
12434 |
+
"step": 1816
|
12435 |
+
},
|
12436 |
+
{
|
12437 |
+
"epoch": 181.7,
|
12438 |
+
"learning_rate": 9.57e-05,
|
12439 |
+
"loss": 0.0048,
|
12440 |
+
"step": 1817
|
12441 |
+
},
|
12442 |
+
{
|
12443 |
+
"epoch": 181.8,
|
12444 |
+
"learning_rate": 9.569473684210527e-05,
|
12445 |
+
"loss": 0.0045,
|
12446 |
+
"step": 1818
|
12447 |
+
},
|
12448 |
+
{
|
12449 |
+
"epoch": 181.9,
|
12450 |
+
"learning_rate": 9.568947368421053e-05,
|
12451 |
+
"loss": 0.0049,
|
12452 |
+
"step": 1819
|
12453 |
+
},
|
12454 |
+
{
|
12455 |
+
"epoch": 182.0,
|
12456 |
+
"learning_rate": 9.56842105263158e-05,
|
12457 |
+
"loss": 0.0065,
|
12458 |
+
"step": 1820
|
12459 |
+
},
|
12460 |
+
{
|
12461 |
+
"epoch": 182.0,
|
12462 |
+
"eval_accuracy_safe": 0.8510607129502199,
|
12463 |
+
"eval_accuracy_unlabeled": NaN,
|
12464 |
+
"eval_accuracy_unsafe": 0.9959411190925398,
|
12465 |
+
"eval_iou_safe": 0.7509239181271631,
|
12466 |
+
"eval_iou_unlabeled": NaN,
|
12467 |
+
"eval_iou_unsafe": 0.9914465595257035,
|
12468 |
+
"eval_loss": 0.04199281707406044,
|
12469 |
+
"eval_mean_accuracy": 0.9235009160213798,
|
12470 |
+
"eval_mean_iou": 0.8711852388264333,
|
12471 |
+
"eval_overall_accuracy": 0.9916615841993645,
|
12472 |
+
"eval_runtime": 9.9031,
|
12473 |
+
"eval_samples_per_second": 6.766,
|
12474 |
+
"eval_steps_per_second": 0.505,
|
12475 |
+
"step": 1820
|
12476 |
+
},
|
12477 |
+
{
|
12478 |
+
"epoch": 182.1,
|
12479 |
+
"learning_rate": 9.567894736842106e-05,
|
12480 |
+
"loss": 0.0054,
|
12481 |
+
"step": 1821
|
12482 |
+
},
|
12483 |
+
{
|
12484 |
+
"epoch": 182.2,
|
12485 |
+
"learning_rate": 9.567368421052632e-05,
|
12486 |
+
"loss": 0.0056,
|
12487 |
+
"step": 1822
|
12488 |
+
},
|
12489 |
+
{
|
12490 |
+
"epoch": 182.3,
|
12491 |
+
"learning_rate": 9.566842105263158e-05,
|
12492 |
+
"loss": 0.0051,
|
12493 |
+
"step": 1823
|
12494 |
+
},
|
12495 |
+
{
|
12496 |
+
"epoch": 182.4,
|
12497 |
+
"learning_rate": 9.566315789473684e-05,
|
12498 |
+
"loss": 0.0057,
|
12499 |
+
"step": 1824
|
12500 |
+
},
|
12501 |
+
{
|
12502 |
+
"epoch": 182.5,
|
12503 |
+
"learning_rate": 9.565789473684211e-05,
|
12504 |
+
"loss": 0.0055,
|
12505 |
+
"step": 1825
|
12506 |
+
},
|
12507 |
+
{
|
12508 |
+
"epoch": 182.6,
|
12509 |
+
"learning_rate": 9.565263157894737e-05,
|
12510 |
+
"loss": 0.0054,
|
12511 |
+
"step": 1826
|
12512 |
+
},
|
12513 |
+
{
|
12514 |
+
"epoch": 182.7,
|
12515 |
+
"learning_rate": 9.564736842105265e-05,
|
12516 |
+
"loss": 0.0054,
|
12517 |
+
"step": 1827
|
12518 |
+
},
|
12519 |
+
{
|
12520 |
+
"epoch": 182.8,
|
12521 |
+
"learning_rate": 9.56421052631579e-05,
|
12522 |
+
"loss": 0.005,
|
12523 |
+
"step": 1828
|
12524 |
+
},
|
12525 |
+
{
|
12526 |
+
"epoch": 182.9,
|
12527 |
+
"learning_rate": 9.563684210526316e-05,
|
12528 |
+
"loss": 0.0053,
|
12529 |
+
"step": 1829
|
12530 |
+
},
|
12531 |
+
{
|
12532 |
+
"epoch": 183.0,
|
12533 |
+
"learning_rate": 9.563157894736843e-05,
|
12534 |
+
"loss": 0.005,
|
12535 |
+
"step": 1830
|
12536 |
+
},
|
12537 |
+
{
|
12538 |
+
"epoch": 183.1,
|
12539 |
+
"learning_rate": 9.562631578947368e-05,
|
12540 |
+
"loss": 0.0052,
|
12541 |
+
"step": 1831
|
12542 |
+
},
|
12543 |
+
{
|
12544 |
+
"epoch": 183.2,
|
12545 |
+
"learning_rate": 9.562105263157896e-05,
|
12546 |
+
"loss": 0.0045,
|
12547 |
+
"step": 1832
|
12548 |
+
},
|
12549 |
+
{
|
12550 |
+
"epoch": 183.3,
|
12551 |
+
"learning_rate": 9.561578947368422e-05,
|
12552 |
+
"loss": 0.0056,
|
12553 |
+
"step": 1833
|
12554 |
+
},
|
12555 |
+
{
|
12556 |
+
"epoch": 183.4,
|
12557 |
+
"learning_rate": 9.561052631578948e-05,
|
12558 |
+
"loss": 0.0047,
|
12559 |
+
"step": 1834
|
12560 |
+
},
|
12561 |
+
{
|
12562 |
+
"epoch": 183.5,
|
12563 |
+
"learning_rate": 9.560526315789474e-05,
|
12564 |
+
"loss": 0.0047,
|
12565 |
+
"step": 1835
|
12566 |
+
},
|
12567 |
+
{
|
12568 |
+
"epoch": 183.6,
|
12569 |
+
"learning_rate": 9.56e-05,
|
12570 |
+
"loss": 0.0053,
|
12571 |
+
"step": 1836
|
12572 |
+
},
|
12573 |
+
{
|
12574 |
+
"epoch": 183.7,
|
12575 |
+
"learning_rate": 9.559473684210527e-05,
|
12576 |
+
"loss": 0.0063,
|
12577 |
+
"step": 1837
|
12578 |
+
},
|
12579 |
+
{
|
12580 |
+
"epoch": 183.8,
|
12581 |
+
"learning_rate": 9.558947368421053e-05,
|
12582 |
+
"loss": 0.005,
|
12583 |
+
"step": 1838
|
12584 |
+
},
|
12585 |
+
{
|
12586 |
+
"epoch": 183.9,
|
12587 |
+
"learning_rate": 9.558421052631579e-05,
|
12588 |
+
"loss": 0.006,
|
12589 |
+
"step": 1839
|
12590 |
+
},
|
12591 |
+
{
|
12592 |
+
"epoch": 184.0,
|
12593 |
+
"learning_rate": 9.557894736842106e-05,
|
12594 |
+
"loss": 0.0045,
|
12595 |
+
"step": 1840
|
12596 |
+
},
|
12597 |
+
{
|
12598 |
+
"epoch": 184.0,
|
12599 |
+
"eval_accuracy_safe": 0.854119683424505,
|
12600 |
+
"eval_accuracy_unlabeled": NaN,
|
12601 |
+
"eval_accuracy_unsafe": 0.9958763487801533,
|
12602 |
+
"eval_iou_safe": 0.7522106167319369,
|
12603 |
+
"eval_iou_unlabeled": NaN,
|
12604 |
+
"eval_iou_unsafe": 0.9914739784033543,
|
12605 |
+
"eval_loss": 0.04210684448480606,
|
12606 |
+
"eval_mean_accuracy": 0.9249980161023291,
|
12607 |
+
"eval_mean_iou": 0.8718422975676456,
|
12608 |
+
"eval_overall_accuracy": 0.9916890841811451,
|
12609 |
+
"eval_runtime": 9.7369,
|
12610 |
+
"eval_samples_per_second": 6.881,
|
12611 |
+
"eval_steps_per_second": 0.514,
|
12612 |
+
"step": 1840
|
12613 |
+
},
|
12614 |
+
{
|
12615 |
+
"epoch": 184.1,
|
12616 |
+
"learning_rate": 9.557368421052632e-05,
|
12617 |
+
"loss": 0.0059,
|
12618 |
+
"step": 1841
|
12619 |
+
},
|
12620 |
+
{
|
12621 |
+
"epoch": 184.2,
|
12622 |
+
"learning_rate": 9.556842105263158e-05,
|
12623 |
+
"loss": 0.0052,
|
12624 |
+
"step": 1842
|
12625 |
+
},
|
12626 |
+
{
|
12627 |
+
"epoch": 184.3,
|
12628 |
+
"learning_rate": 9.556315789473684e-05,
|
12629 |
+
"loss": 0.0048,
|
12630 |
+
"step": 1843
|
12631 |
+
},
|
12632 |
+
{
|
12633 |
+
"epoch": 184.4,
|
12634 |
+
"learning_rate": 9.555789473684212e-05,
|
12635 |
+
"loss": 0.0049,
|
12636 |
+
"step": 1844
|
12637 |
+
},
|
12638 |
+
{
|
12639 |
+
"epoch": 184.5,
|
12640 |
+
"learning_rate": 9.555263157894737e-05,
|
12641 |
+
"loss": 0.0044,
|
12642 |
+
"step": 1845
|
12643 |
+
},
|
12644 |
+
{
|
12645 |
+
"epoch": 184.6,
|
12646 |
+
"learning_rate": 9.554736842105264e-05,
|
12647 |
+
"loss": 0.0056,
|
12648 |
+
"step": 1846
|
12649 |
+
},
|
12650 |
+
{
|
12651 |
+
"epoch": 184.7,
|
12652 |
+
"learning_rate": 9.55421052631579e-05,
|
12653 |
+
"loss": 0.0057,
|
12654 |
+
"step": 1847
|
12655 |
+
},
|
12656 |
+
{
|
12657 |
+
"epoch": 184.8,
|
12658 |
+
"learning_rate": 9.553684210526315e-05,
|
12659 |
+
"loss": 0.0055,
|
12660 |
+
"step": 1848
|
12661 |
+
},
|
12662 |
+
{
|
12663 |
+
"epoch": 184.9,
|
12664 |
+
"learning_rate": 9.553157894736843e-05,
|
12665 |
+
"loss": 0.0044,
|
12666 |
+
"step": 1849
|
12667 |
+
},
|
12668 |
+
{
|
12669 |
+
"epoch": 185.0,
|
12670 |
+
"learning_rate": 9.552631578947369e-05,
|
12671 |
+
"loss": 0.005,
|
12672 |
+
"step": 1850
|
12673 |
+
},
|
12674 |
+
{
|
12675 |
+
"epoch": 185.1,
|
12676 |
+
"learning_rate": 9.552105263157895e-05,
|
12677 |
+
"loss": 0.0057,
|
12678 |
+
"step": 1851
|
12679 |
+
},
|
12680 |
+
{
|
12681 |
+
"epoch": 185.2,
|
12682 |
+
"learning_rate": 9.551578947368422e-05,
|
12683 |
+
"loss": 0.006,
|
12684 |
+
"step": 1852
|
12685 |
+
},
|
12686 |
+
{
|
12687 |
+
"epoch": 185.3,
|
12688 |
+
"learning_rate": 9.551052631578948e-05,
|
12689 |
+
"loss": 0.0047,
|
12690 |
+
"step": 1853
|
12691 |
+
},
|
12692 |
+
{
|
12693 |
+
"epoch": 185.4,
|
12694 |
+
"learning_rate": 9.550526315789474e-05,
|
12695 |
+
"loss": 0.0055,
|
12696 |
+
"step": 1854
|
12697 |
+
},
|
12698 |
+
{
|
12699 |
+
"epoch": 185.5,
|
12700 |
+
"learning_rate": 9.55e-05,
|
12701 |
+
"loss": 0.0049,
|
12702 |
+
"step": 1855
|
12703 |
+
},
|
12704 |
+
{
|
12705 |
+
"epoch": 185.6,
|
12706 |
+
"learning_rate": 9.549473684210527e-05,
|
12707 |
+
"loss": 0.0056,
|
12708 |
+
"step": 1856
|
12709 |
+
},
|
12710 |
+
{
|
12711 |
+
"epoch": 185.7,
|
12712 |
+
"learning_rate": 9.548947368421053e-05,
|
12713 |
+
"loss": 0.0049,
|
12714 |
+
"step": 1857
|
12715 |
+
},
|
12716 |
+
{
|
12717 |
+
"epoch": 185.8,
|
12718 |
+
"learning_rate": 9.54842105263158e-05,
|
12719 |
+
"loss": 0.0041,
|
12720 |
+
"step": 1858
|
12721 |
+
},
|
12722 |
+
{
|
12723 |
+
"epoch": 185.9,
|
12724 |
+
"learning_rate": 9.547894736842105e-05,
|
12725 |
+
"loss": 0.0051,
|
12726 |
+
"step": 1859
|
12727 |
+
},
|
12728 |
+
{
|
12729 |
+
"epoch": 186.0,
|
12730 |
+
"learning_rate": 9.547368421052632e-05,
|
12731 |
+
"loss": 0.0056,
|
12732 |
+
"step": 1860
|
12733 |
+
},
|
12734 |
+
{
|
12735 |
+
"epoch": 186.0,
|
12736 |
+
"eval_accuracy_safe": 0.834923535375731,
|
12737 |
+
"eval_accuracy_unlabeled": NaN,
|
12738 |
+
"eval_accuracy_unsafe": 0.9965099127325644,
|
12739 |
+
"eval_iou_safe": 0.7490359508205225,
|
12740 |
+
"eval_iou_unlabeled": NaN,
|
12741 |
+
"eval_iou_unsafe": 0.9915279704816349,
|
12742 |
+
"eval_loss": 0.043502531945705414,
|
12743 |
+
"eval_mean_accuracy": 0.9157167240541477,
|
12744 |
+
"eval_mean_iou": 0.8702819606510788,
|
12745 |
+
"eval_overall_accuracy": 0.9917369102364155,
|
12746 |
+
"eval_runtime": 9.8978,
|
12747 |
+
"eval_samples_per_second": 6.769,
|
12748 |
+
"eval_steps_per_second": 0.505,
|
12749 |
+
"step": 1860
|
12750 |
+
},
|
12751 |
+
{
|
12752 |
+
"epoch": 186.1,
|
12753 |
+
"learning_rate": 9.546842105263159e-05,
|
12754 |
+
"loss": 0.0048,
|
12755 |
+
"step": 1861
|
12756 |
+
},
|
12757 |
+
{
|
12758 |
+
"epoch": 186.2,
|
12759 |
+
"learning_rate": 9.546315789473684e-05,
|
12760 |
+
"loss": 0.0046,
|
12761 |
+
"step": 1862
|
12762 |
+
},
|
12763 |
+
{
|
12764 |
+
"epoch": 186.3,
|
12765 |
+
"learning_rate": 9.545789473684212e-05,
|
12766 |
+
"loss": 0.0043,
|
12767 |
+
"step": 1863
|
12768 |
+
},
|
12769 |
+
{
|
12770 |
+
"epoch": 186.4,
|
12771 |
+
"learning_rate": 9.545263157894736e-05,
|
12772 |
+
"loss": 0.0055,
|
12773 |
+
"step": 1864
|
12774 |
+
},
|
12775 |
+
{
|
12776 |
+
"epoch": 186.5,
|
12777 |
+
"learning_rate": 9.544736842105264e-05,
|
12778 |
+
"loss": 0.0054,
|
12779 |
+
"step": 1865
|
12780 |
+
},
|
12781 |
+
{
|
12782 |
+
"epoch": 186.6,
|
12783 |
+
"learning_rate": 9.54421052631579e-05,
|
12784 |
+
"loss": 0.0048,
|
12785 |
+
"step": 1866
|
12786 |
+
},
|
12787 |
+
{
|
12788 |
+
"epoch": 186.7,
|
12789 |
+
"learning_rate": 9.543684210526315e-05,
|
12790 |
+
"loss": 0.0053,
|
12791 |
+
"step": 1867
|
12792 |
+
},
|
12793 |
+
{
|
12794 |
+
"epoch": 186.8,
|
12795 |
+
"learning_rate": 9.543157894736843e-05,
|
12796 |
+
"loss": 0.005,
|
12797 |
+
"step": 1868
|
12798 |
+
},
|
12799 |
+
{
|
12800 |
+
"epoch": 186.9,
|
12801 |
+
"learning_rate": 9.542631578947369e-05,
|
12802 |
+
"loss": 0.0057,
|
12803 |
+
"step": 1869
|
12804 |
+
},
|
12805 |
+
{
|
12806 |
+
"epoch": 187.0,
|
12807 |
+
"learning_rate": 9.542105263157895e-05,
|
12808 |
+
"loss": 0.0055,
|
12809 |
+
"step": 1870
|
12810 |
+
},
|
12811 |
+
{
|
12812 |
+
"epoch": 187.1,
|
12813 |
+
"learning_rate": 9.541578947368421e-05,
|
12814 |
+
"loss": 0.0056,
|
12815 |
+
"step": 1871
|
12816 |
+
},
|
12817 |
+
{
|
12818 |
+
"epoch": 187.2,
|
12819 |
+
"learning_rate": 9.541052631578948e-05,
|
12820 |
+
"loss": 0.0052,
|
12821 |
+
"step": 1872
|
12822 |
+
},
|
12823 |
+
{
|
12824 |
+
"epoch": 187.3,
|
12825 |
+
"learning_rate": 9.540526315789474e-05,
|
12826 |
+
"loss": 0.0043,
|
12827 |
+
"step": 1873
|
12828 |
+
},
|
12829 |
+
{
|
12830 |
+
"epoch": 187.4,
|
12831 |
+
"learning_rate": 9.54e-05,
|
12832 |
+
"loss": 0.0042,
|
12833 |
+
"step": 1874
|
12834 |
+
},
|
12835 |
+
{
|
12836 |
+
"epoch": 187.5,
|
12837 |
+
"learning_rate": 9.539473684210526e-05,
|
12838 |
+
"loss": 0.0054,
|
12839 |
+
"step": 1875
|
12840 |
+
},
|
12841 |
+
{
|
12842 |
+
"epoch": 187.6,
|
12843 |
+
"learning_rate": 9.538947368421053e-05,
|
12844 |
+
"loss": 0.0055,
|
12845 |
+
"step": 1876
|
12846 |
+
},
|
12847 |
+
{
|
12848 |
+
"epoch": 187.7,
|
12849 |
+
"learning_rate": 9.53842105263158e-05,
|
12850 |
+
"loss": 0.0051,
|
12851 |
+
"step": 1877
|
12852 |
+
},
|
12853 |
+
{
|
12854 |
+
"epoch": 187.8,
|
12855 |
+
"learning_rate": 9.537894736842105e-05,
|
12856 |
+
"loss": 0.0049,
|
12857 |
+
"step": 1878
|
12858 |
+
},
|
12859 |
+
{
|
12860 |
+
"epoch": 187.9,
|
12861 |
+
"learning_rate": 9.537368421052631e-05,
|
12862 |
+
"loss": 0.0052,
|
12863 |
+
"step": 1879
|
12864 |
+
},
|
12865 |
+
{
|
12866 |
+
"epoch": 188.0,
|
12867 |
+
"learning_rate": 9.536842105263159e-05,
|
12868 |
+
"loss": 0.0059,
|
12869 |
+
"step": 1880
|
12870 |
+
},
|
12871 |
+
{
|
12872 |
+
"epoch": 188.0,
|
12873 |
+
"eval_accuracy_safe": 0.841870308903975,
|
12874 |
+
"eval_accuracy_unlabeled": NaN,
|
12875 |
+
"eval_accuracy_unsafe": 0.9962683734426231,
|
12876 |
+
"eval_iou_safe": 0.7499291732413931,
|
12877 |
+
"eval_iou_unlabeled": NaN,
|
12878 |
+
"eval_iou_unsafe": 0.9914962348083867,
|
12879 |
+
"eval_loss": 0.04357173666357994,
|
12880 |
+
"eval_mean_accuracy": 0.9190693411732991,
|
12881 |
+
"eval_mean_iou": 0.8707127040248899,
|
12882 |
+
"eval_overall_accuracy": 0.9917077021812325,
|
12883 |
+
"eval_runtime": 9.8486,
|
12884 |
+
"eval_samples_per_second": 6.803,
|
12885 |
+
"eval_steps_per_second": 0.508,
|
12886 |
+
"step": 1880
|
12887 |
+
},
|
12888 |
+
{
|
12889 |
+
"epoch": 188.1,
|
12890 |
+
"learning_rate": 9.536315789473684e-05,
|
12891 |
+
"loss": 0.0051,
|
12892 |
+
"step": 1881
|
12893 |
+
},
|
12894 |
+
{
|
12895 |
+
"epoch": 188.2,
|
12896 |
+
"learning_rate": 9.535789473684211e-05,
|
12897 |
+
"loss": 0.0047,
|
12898 |
+
"step": 1882
|
12899 |
+
},
|
12900 |
+
{
|
12901 |
+
"epoch": 188.3,
|
12902 |
+
"learning_rate": 9.535263157894738e-05,
|
12903 |
+
"loss": 0.0054,
|
12904 |
+
"step": 1883
|
12905 |
+
},
|
12906 |
+
{
|
12907 |
+
"epoch": 188.4,
|
12908 |
+
"learning_rate": 9.534736842105264e-05,
|
12909 |
+
"loss": 0.0058,
|
12910 |
+
"step": 1884
|
12911 |
+
},
|
12912 |
+
{
|
12913 |
+
"epoch": 188.5,
|
12914 |
+
"learning_rate": 9.53421052631579e-05,
|
12915 |
+
"loss": 0.0046,
|
12916 |
+
"step": 1885
|
12917 |
+
},
|
12918 |
+
{
|
12919 |
+
"epoch": 188.6,
|
12920 |
+
"learning_rate": 9.533684210526316e-05,
|
12921 |
+
"loss": 0.0049,
|
12922 |
+
"step": 1886
|
12923 |
+
},
|
12924 |
+
{
|
12925 |
+
"epoch": 188.7,
|
12926 |
+
"learning_rate": 9.533157894736843e-05,
|
12927 |
+
"loss": 0.0055,
|
12928 |
+
"step": 1887
|
12929 |
+
},
|
12930 |
+
{
|
12931 |
+
"epoch": 188.8,
|
12932 |
+
"learning_rate": 9.532631578947369e-05,
|
12933 |
+
"loss": 0.0044,
|
12934 |
+
"step": 1888
|
12935 |
+
},
|
12936 |
+
{
|
12937 |
+
"epoch": 188.9,
|
12938 |
+
"learning_rate": 9.532105263157895e-05,
|
12939 |
+
"loss": 0.0059,
|
12940 |
+
"step": 1889
|
12941 |
+
},
|
12942 |
+
{
|
12943 |
+
"epoch": 189.0,
|
12944 |
+
"learning_rate": 9.531578947368421e-05,
|
12945 |
+
"loss": 0.0054,
|
12946 |
+
"step": 1890
|
12947 |
+
},
|
12948 |
+
{
|
12949 |
+
"epoch": 189.1,
|
12950 |
+
"learning_rate": 9.531052631578949e-05,
|
12951 |
+
"loss": 0.0051,
|
12952 |
+
"step": 1891
|
12953 |
+
},
|
12954 |
+
{
|
12955 |
+
"epoch": 189.2,
|
12956 |
+
"learning_rate": 9.530526315789474e-05,
|
12957 |
+
"loss": 0.0054,
|
12958 |
+
"step": 1892
|
12959 |
+
},
|
12960 |
+
{
|
12961 |
+
"epoch": 189.3,
|
12962 |
+
"learning_rate": 9.53e-05,
|
12963 |
+
"loss": 0.0049,
|
12964 |
+
"step": 1893
|
12965 |
+
},
|
12966 |
+
{
|
12967 |
+
"epoch": 189.4,
|
12968 |
+
"learning_rate": 9.529473684210527e-05,
|
12969 |
+
"loss": 0.0052,
|
12970 |
+
"step": 1894
|
12971 |
+
},
|
12972 |
+
{
|
12973 |
+
"epoch": 189.5,
|
12974 |
+
"learning_rate": 9.528947368421052e-05,
|
12975 |
+
"loss": 0.0053,
|
12976 |
+
"step": 1895
|
12977 |
+
},
|
12978 |
+
{
|
12979 |
+
"epoch": 189.6,
|
12980 |
+
"learning_rate": 9.52842105263158e-05,
|
12981 |
+
"loss": 0.0071,
|
12982 |
+
"step": 1896
|
12983 |
+
},
|
12984 |
+
{
|
12985 |
+
"epoch": 189.7,
|
12986 |
+
"learning_rate": 9.527894736842106e-05,
|
12987 |
+
"loss": 0.0052,
|
12988 |
+
"step": 1897
|
12989 |
+
},
|
12990 |
+
{
|
12991 |
+
"epoch": 189.8,
|
12992 |
+
"learning_rate": 9.527368421052631e-05,
|
12993 |
+
"loss": 0.0043,
|
12994 |
+
"step": 1898
|
12995 |
+
},
|
12996 |
+
{
|
12997 |
+
"epoch": 189.9,
|
12998 |
+
"learning_rate": 9.526842105263159e-05,
|
12999 |
+
"loss": 0.0044,
|
13000 |
+
"step": 1899
|
13001 |
+
},
|
13002 |
+
{
|
13003 |
+
"epoch": 190.0,
|
13004 |
+
"learning_rate": 9.526315789473685e-05,
|
13005 |
+
"loss": 0.0042,
|
13006 |
+
"step": 1900
|
13007 |
+
},
|
13008 |
+
{
|
13009 |
+
"epoch": 190.0,
|
13010 |
+
"eval_accuracy_safe": 0.8458140099691211,
|
13011 |
+
"eval_accuracy_unlabeled": NaN,
|
13012 |
+
"eval_accuracy_unsafe": 0.9961090877559117,
|
13013 |
+
"eval_iou_safe": 0.7499461650732667,
|
13014 |
+
"eval_iou_unlabeled": NaN,
|
13015 |
+
"eval_iou_unsafe": 0.9914561527530946,
|
13016 |
+
"eval_loss": 0.04357015714049339,
|
13017 |
+
"eval_mean_accuracy": 0.9209615488625165,
|
13018 |
+
"eval_mean_iou": 0.8707011589131807,
|
13019 |
+
"eval_overall_accuracy": 0.9916696121443563,
|
13020 |
+
"eval_runtime": 9.6908,
|
13021 |
+
"eval_samples_per_second": 6.914,
|
13022 |
+
"eval_steps_per_second": 0.516,
|
13023 |
+
"step": 1900
|
13024 |
+
},
|
13025 |
+
{
|
13026 |
+
"epoch": 190.1,
|
13027 |
+
"learning_rate": 9.525789473684211e-05,
|
13028 |
+
"loss": 0.0046,
|
13029 |
+
"step": 1901
|
13030 |
+
},
|
13031 |
+
{
|
13032 |
+
"epoch": 190.2,
|
13033 |
+
"learning_rate": 9.525263157894737e-05,
|
13034 |
+
"loss": 0.0052,
|
13035 |
+
"step": 1902
|
13036 |
+
},
|
13037 |
+
{
|
13038 |
+
"epoch": 190.3,
|
13039 |
+
"learning_rate": 9.524736842105264e-05,
|
13040 |
+
"loss": 0.005,
|
13041 |
+
"step": 1903
|
13042 |
+
},
|
13043 |
+
{
|
13044 |
+
"epoch": 190.4,
|
13045 |
+
"learning_rate": 9.52421052631579e-05,
|
13046 |
+
"loss": 0.0052,
|
13047 |
+
"step": 1904
|
13048 |
+
},
|
13049 |
+
{
|
13050 |
+
"epoch": 190.5,
|
13051 |
+
"learning_rate": 9.523684210526316e-05,
|
13052 |
+
"loss": 0.0048,
|
13053 |
+
"step": 1905
|
13054 |
+
},
|
13055 |
+
{
|
13056 |
+
"epoch": 190.6,
|
13057 |
+
"learning_rate": 9.523157894736842e-05,
|
13058 |
+
"loss": 0.005,
|
13059 |
+
"step": 1906
|
13060 |
+
},
|
13061 |
+
{
|
13062 |
+
"epoch": 190.7,
|
13063 |
+
"learning_rate": 9.522631578947368e-05,
|
13064 |
+
"loss": 0.0053,
|
13065 |
+
"step": 1907
|
13066 |
+
},
|
13067 |
+
{
|
13068 |
+
"epoch": 190.8,
|
13069 |
+
"learning_rate": 9.522105263157896e-05,
|
13070 |
+
"loss": 0.0065,
|
13071 |
+
"step": 1908
|
13072 |
+
},
|
13073 |
+
{
|
13074 |
+
"epoch": 190.9,
|
13075 |
+
"learning_rate": 9.521578947368421e-05,
|
13076 |
+
"loss": 0.0056,
|
13077 |
+
"step": 1909
|
13078 |
+
},
|
13079 |
+
{
|
13080 |
+
"epoch": 191.0,
|
13081 |
+
"learning_rate": 9.521052631578949e-05,
|
13082 |
+
"loss": 0.0061,
|
13083 |
+
"step": 1910
|
13084 |
+
},
|
13085 |
+
{
|
13086 |
+
"epoch": 191.1,
|
13087 |
+
"learning_rate": 9.520526315789475e-05,
|
13088 |
+
"loss": 0.0061,
|
13089 |
+
"step": 1911
|
13090 |
+
},
|
13091 |
+
{
|
13092 |
+
"epoch": 191.2,
|
13093 |
+
"learning_rate": 9.52e-05,
|
13094 |
+
"loss": 0.0044,
|
13095 |
+
"step": 1912
|
13096 |
+
},
|
13097 |
+
{
|
13098 |
+
"epoch": 191.3,
|
13099 |
+
"learning_rate": 9.519473684210527e-05,
|
13100 |
+
"loss": 0.0056,
|
13101 |
+
"step": 1913
|
13102 |
+
},
|
13103 |
+
{
|
13104 |
+
"epoch": 191.4,
|
13105 |
+
"learning_rate": 9.518947368421053e-05,
|
13106 |
+
"loss": 0.005,
|
13107 |
+
"step": 1914
|
13108 |
+
},
|
13109 |
+
{
|
13110 |
+
"epoch": 191.5,
|
13111 |
+
"learning_rate": 9.51842105263158e-05,
|
13112 |
+
"loss": 0.0046,
|
13113 |
+
"step": 1915
|
13114 |
+
},
|
13115 |
+
{
|
13116 |
+
"epoch": 191.6,
|
13117 |
+
"learning_rate": 9.517894736842106e-05,
|
13118 |
+
"loss": 0.0054,
|
13119 |
+
"step": 1916
|
13120 |
+
},
|
13121 |
+
{
|
13122 |
+
"epoch": 191.7,
|
13123 |
+
"learning_rate": 9.517368421052631e-05,
|
13124 |
+
"loss": 0.0044,
|
13125 |
+
"step": 1917
|
13126 |
+
},
|
13127 |
+
{
|
13128 |
+
"epoch": 191.8,
|
13129 |
+
"learning_rate": 9.516842105263158e-05,
|
13130 |
+
"loss": 0.0048,
|
13131 |
+
"step": 1918
|
13132 |
+
},
|
13133 |
+
{
|
13134 |
+
"epoch": 191.9,
|
13135 |
+
"learning_rate": 9.516315789473685e-05,
|
13136 |
+
"loss": 0.006,
|
13137 |
+
"step": 1919
|
13138 |
+
},
|
13139 |
+
{
|
13140 |
+
"epoch": 192.0,
|
13141 |
+
"learning_rate": 9.515789473684211e-05,
|
13142 |
+
"loss": 0.006,
|
13143 |
+
"step": 1920
|
13144 |
+
},
|
13145 |
+
{
|
13146 |
+
"epoch": 192.0,
|
13147 |
+
"eval_accuracy_safe": 0.8425140997914425,
|
13148 |
+
"eval_accuracy_unlabeled": NaN,
|
13149 |
+
"eval_accuracy_unsafe": 0.996152971989304,
|
13150 |
+
"eval_iou_safe": 0.7479764671255053,
|
13151 |
+
"eval_iou_unlabeled": NaN,
|
13152 |
+
"eval_iou_unsafe": 0.991400719934838,
|
13153 |
+
"eval_loss": 0.042638879269361496,
|
13154 |
+
"eval_mean_accuracy": 0.9193335358903733,
|
13155 |
+
"eval_mean_iou": 0.8696885935301717,
|
13156 |
+
"eval_overall_accuracy": 0.9916147260523555,
|
13157 |
+
"eval_runtime": 10.1388,
|
13158 |
+
"eval_samples_per_second": 6.608,
|
13159 |
+
"eval_steps_per_second": 0.493,
|
13160 |
+
"step": 1920
|
13161 |
+
},
|
13162 |
+
{
|
13163 |
+
"epoch": 192.1,
|
13164 |
+
"learning_rate": 9.515263157894737e-05,
|
13165 |
+
"loss": 0.0049,
|
13166 |
+
"step": 1921
|
13167 |
+
},
|
13168 |
+
{
|
13169 |
+
"epoch": 192.2,
|
13170 |
+
"learning_rate": 9.514736842105265e-05,
|
13171 |
+
"loss": 0.0049,
|
13172 |
+
"step": 1922
|
13173 |
+
},
|
13174 |
+
{
|
13175 |
+
"epoch": 192.3,
|
13176 |
+
"learning_rate": 9.51421052631579e-05,
|
13177 |
+
"loss": 0.0053,
|
13178 |
+
"step": 1923
|
13179 |
+
},
|
13180 |
+
{
|
13181 |
+
"epoch": 192.4,
|
13182 |
+
"learning_rate": 9.513684210526316e-05,
|
13183 |
+
"loss": 0.0043,
|
13184 |
+
"step": 1924
|
13185 |
+
},
|
13186 |
+
{
|
13187 |
+
"epoch": 192.5,
|
13188 |
+
"learning_rate": 9.513157894736843e-05,
|
13189 |
+
"loss": 0.0067,
|
13190 |
+
"step": 1925
|
13191 |
+
},
|
13192 |
+
{
|
13193 |
+
"epoch": 192.6,
|
13194 |
+
"learning_rate": 9.512631578947368e-05,
|
13195 |
+
"loss": 0.0045,
|
13196 |
+
"step": 1926
|
13197 |
+
},
|
13198 |
+
{
|
13199 |
+
"epoch": 192.7,
|
13200 |
+
"learning_rate": 9.512105263157896e-05,
|
13201 |
+
"loss": 0.0059,
|
13202 |
+
"step": 1927
|
13203 |
+
},
|
13204 |
+
{
|
13205 |
+
"epoch": 192.8,
|
13206 |
+
"learning_rate": 9.511578947368421e-05,
|
13207 |
+
"loss": 0.0043,
|
13208 |
+
"step": 1928
|
13209 |
+
},
|
13210 |
+
{
|
13211 |
+
"epoch": 192.9,
|
13212 |
+
"learning_rate": 9.511052631578948e-05,
|
13213 |
+
"loss": 0.0048,
|
13214 |
+
"step": 1929
|
13215 |
+
},
|
13216 |
+
{
|
13217 |
+
"epoch": 193.0,
|
13218 |
+
"learning_rate": 9.510526315789475e-05,
|
13219 |
+
"loss": 0.0055,
|
13220 |
+
"step": 1930
|
13221 |
+
},
|
13222 |
+
{
|
13223 |
+
"epoch": 193.1,
|
13224 |
+
"learning_rate": 9.51e-05,
|
13225 |
+
"loss": 0.0046,
|
13226 |
+
"step": 1931
|
13227 |
+
},
|
13228 |
+
{
|
13229 |
+
"epoch": 193.2,
|
13230 |
+
"learning_rate": 9.509473684210527e-05,
|
13231 |
+
"loss": 0.0048,
|
13232 |
+
"step": 1932
|
13233 |
+
},
|
13234 |
+
{
|
13235 |
+
"epoch": 193.3,
|
13236 |
+
"learning_rate": 9.508947368421053e-05,
|
13237 |
+
"loss": 0.0055,
|
13238 |
+
"step": 1933
|
13239 |
+
},
|
13240 |
+
{
|
13241 |
+
"epoch": 193.4,
|
13242 |
+
"learning_rate": 9.50842105263158e-05,
|
13243 |
+
"loss": 0.005,
|
13244 |
+
"step": 1934
|
13245 |
+
},
|
13246 |
+
{
|
13247 |
+
"epoch": 193.5,
|
13248 |
+
"learning_rate": 9.507894736842106e-05,
|
13249 |
+
"loss": 0.0047,
|
13250 |
+
"step": 1935
|
13251 |
+
},
|
13252 |
+
{
|
13253 |
+
"epoch": 193.6,
|
13254 |
+
"learning_rate": 9.507368421052632e-05,
|
13255 |
+
"loss": 0.0052,
|
13256 |
+
"step": 1936
|
13257 |
+
},
|
13258 |
+
{
|
13259 |
+
"epoch": 193.7,
|
13260 |
+
"learning_rate": 9.506842105263158e-05,
|
13261 |
+
"loss": 0.0066,
|
13262 |
+
"step": 1937
|
13263 |
+
},
|
13264 |
+
{
|
13265 |
+
"epoch": 193.8,
|
13266 |
+
"learning_rate": 9.506315789473684e-05,
|
13267 |
+
"loss": 0.0044,
|
13268 |
+
"step": 1938
|
13269 |
+
},
|
13270 |
+
{
|
13271 |
+
"epoch": 193.9,
|
13272 |
+
"learning_rate": 9.50578947368421e-05,
|
13273 |
+
"loss": 0.0054,
|
13274 |
+
"step": 1939
|
13275 |
+
},
|
13276 |
+
{
|
13277 |
+
"epoch": 194.0,
|
13278 |
+
"learning_rate": 9.505263157894737e-05,
|
13279 |
+
"loss": 0.0053,
|
13280 |
+
"step": 1940
|
13281 |
+
},
|
13282 |
+
{
|
13283 |
+
"epoch": 194.0,
|
13284 |
+
"eval_accuracy_safe": 0.843699523132139,
|
13285 |
+
"eval_accuracy_unlabeled": NaN,
|
13286 |
+
"eval_accuracy_unsafe": 0.9961060956490895,
|
13287 |
+
"eval_iou_safe": 0.7480061451912511,
|
13288 |
+
"eval_iou_unlabeled": NaN,
|
13289 |
+
"eval_iou_unsafe": 0.9913896671918935,
|
13290 |
+
"eval_loss": 0.04471457004547119,
|
13291 |
+
"eval_mean_accuracy": 0.9199028093906143,
|
13292 |
+
"eval_mean_iou": 0.8696979061915723,
|
13293 |
+
"eval_overall_accuracy": 0.99160424986882,
|
13294 |
+
"eval_runtime": 10.043,
|
13295 |
+
"eval_samples_per_second": 6.671,
|
13296 |
+
"eval_steps_per_second": 0.498,
|
13297 |
+
"step": 1940
|
13298 |
+
},
|
13299 |
+
{
|
13300 |
+
"epoch": 194.1,
|
13301 |
+
"learning_rate": 9.504736842105264e-05,
|
13302 |
+
"loss": 0.0056,
|
13303 |
+
"step": 1941
|
13304 |
+
},
|
13305 |
+
{
|
13306 |
+
"epoch": 194.2,
|
13307 |
+
"learning_rate": 9.504210526315789e-05,
|
13308 |
+
"loss": 0.0048,
|
13309 |
+
"step": 1942
|
13310 |
+
},
|
13311 |
+
{
|
13312 |
+
"epoch": 194.3,
|
13313 |
+
"learning_rate": 9.503684210526316e-05,
|
13314 |
+
"loss": 0.0041,
|
13315 |
+
"step": 1943
|
13316 |
+
},
|
13317 |
+
{
|
13318 |
+
"epoch": 194.4,
|
13319 |
+
"learning_rate": 9.503157894736843e-05,
|
13320 |
+
"loss": 0.0042,
|
13321 |
+
"step": 1944
|
13322 |
+
},
|
13323 |
+
{
|
13324 |
+
"epoch": 194.5,
|
13325 |
+
"learning_rate": 9.502631578947368e-05,
|
13326 |
+
"loss": 0.0041,
|
13327 |
+
"step": 1945
|
13328 |
+
},
|
13329 |
+
{
|
13330 |
+
"epoch": 194.6,
|
13331 |
+
"learning_rate": 9.502105263157896e-05,
|
13332 |
+
"loss": 0.0058,
|
13333 |
+
"step": 1946
|
13334 |
+
},
|
13335 |
+
{
|
13336 |
+
"epoch": 194.7,
|
13337 |
+
"learning_rate": 9.501578947368422e-05,
|
13338 |
+
"loss": 0.0057,
|
13339 |
+
"step": 1947
|
13340 |
+
},
|
13341 |
+
{
|
13342 |
+
"epoch": 194.8,
|
13343 |
+
"learning_rate": 9.501052631578948e-05,
|
13344 |
+
"loss": 0.004,
|
13345 |
+
"step": 1948
|
13346 |
+
},
|
13347 |
+
{
|
13348 |
+
"epoch": 194.9,
|
13349 |
+
"learning_rate": 9.500526315789474e-05,
|
13350 |
+
"loss": 0.0053,
|
13351 |
+
"step": 1949
|
13352 |
+
},
|
13353 |
+
{
|
13354 |
+
"epoch": 195.0,
|
13355 |
+
"learning_rate": 9.5e-05,
|
13356 |
+
"loss": 0.0061,
|
13357 |
+
"step": 1950
|
13358 |
+
},
|
13359 |
+
{
|
13360 |
+
"epoch": 195.1,
|
13361 |
+
"learning_rate": 9.499473684210527e-05,
|
13362 |
+
"loss": 0.0058,
|
13363 |
+
"step": 1951
|
13364 |
+
},
|
13365 |
+
{
|
13366 |
+
"epoch": 195.2,
|
13367 |
+
"learning_rate": 9.498947368421053e-05,
|
13368 |
+
"loss": 0.0049,
|
13369 |
+
"step": 1952
|
13370 |
+
},
|
13371 |
+
{
|
13372 |
+
"epoch": 195.3,
|
13373 |
+
"learning_rate": 9.498421052631579e-05,
|
13374 |
+
"loss": 0.0055,
|
13375 |
+
"step": 1953
|
13376 |
+
},
|
13377 |
+
{
|
13378 |
+
"epoch": 195.4,
|
13379 |
+
"learning_rate": 9.497894736842105e-05,
|
13380 |
+
"loss": 0.0051,
|
13381 |
+
"step": 1954
|
13382 |
+
},
|
13383 |
+
{
|
13384 |
+
"epoch": 195.5,
|
13385 |
+
"learning_rate": 9.497368421052633e-05,
|
13386 |
+
"loss": 0.005,
|
13387 |
+
"step": 1955
|
13388 |
+
},
|
13389 |
+
{
|
13390 |
+
"epoch": 195.6,
|
13391 |
+
"learning_rate": 9.496842105263158e-05,
|
13392 |
+
"loss": 0.0048,
|
13393 |
+
"step": 1956
|
13394 |
+
},
|
13395 |
+
{
|
13396 |
+
"epoch": 195.7,
|
13397 |
+
"learning_rate": 9.496315789473684e-05,
|
13398 |
+
"loss": 0.0038,
|
13399 |
+
"step": 1957
|
13400 |
+
},
|
13401 |
+
{
|
13402 |
+
"epoch": 195.8,
|
13403 |
+
"learning_rate": 9.495789473684212e-05,
|
13404 |
+
"loss": 0.0041,
|
13405 |
+
"step": 1958
|
13406 |
+
},
|
13407 |
+
{
|
13408 |
+
"epoch": 195.9,
|
13409 |
+
"learning_rate": 9.495263157894737e-05,
|
13410 |
+
"loss": 0.0047,
|
13411 |
+
"step": 1959
|
13412 |
+
},
|
13413 |
+
{
|
13414 |
+
"epoch": 196.0,
|
13415 |
+
"learning_rate": 9.494736842105264e-05,
|
13416 |
+
"loss": 0.0044,
|
13417 |
+
"step": 1960
|
13418 |
+
},
|
13419 |
+
{
|
13420 |
+
"epoch": 196.0,
|
13421 |
+
"eval_accuracy_safe": 0.8516408957559917,
|
13422 |
+
"eval_accuracy_unlabeled": NaN,
|
13423 |
+
"eval_accuracy_unsafe": 0.9958981148905658,
|
13424 |
+
"eval_iou_safe": 0.7505002395034635,
|
13425 |
+
"eval_iou_unlabeled": NaN,
|
13426 |
+
"eval_iou_unsafe": 0.9914211781870088,
|
13427 |
+
"eval_loss": 0.04413814842700958,
|
13428 |
+
"eval_mean_accuracy": 0.9237695053232787,
|
13429 |
+
"eval_mean_iou": 0.8709607088452361,
|
13430 |
+
"eval_overall_accuracy": 0.9916369879423682,
|
13431 |
+
"eval_runtime": 9.6827,
|
13432 |
+
"eval_samples_per_second": 6.92,
|
13433 |
+
"eval_steps_per_second": 0.516,
|
13434 |
+
"step": 1960
|
13435 |
+
},
|
13436 |
+
{
|
13437 |
+
"epoch": 196.1,
|
13438 |
+
"learning_rate": 9.49421052631579e-05,
|
13439 |
+
"loss": 0.0049,
|
13440 |
+
"step": 1961
|
13441 |
+
},
|
13442 |
+
{
|
13443 |
+
"epoch": 196.2,
|
13444 |
+
"learning_rate": 9.493684210526315e-05,
|
13445 |
+
"loss": 0.0055,
|
13446 |
+
"step": 1962
|
13447 |
+
},
|
13448 |
+
{
|
13449 |
+
"epoch": 196.3,
|
13450 |
+
"learning_rate": 9.493157894736843e-05,
|
13451 |
+
"loss": 0.0048,
|
13452 |
+
"step": 1963
|
13453 |
+
},
|
13454 |
+
{
|
13455 |
+
"epoch": 196.4,
|
13456 |
+
"learning_rate": 9.492631578947369e-05,
|
13457 |
+
"loss": 0.004,
|
13458 |
+
"step": 1964
|
13459 |
+
},
|
13460 |
+
{
|
13461 |
+
"epoch": 196.5,
|
13462 |
+
"learning_rate": 9.492105263157895e-05,
|
13463 |
+
"loss": 0.0049,
|
13464 |
+
"step": 1965
|
13465 |
+
},
|
13466 |
+
{
|
13467 |
+
"epoch": 196.6,
|
13468 |
+
"learning_rate": 9.491578947368422e-05,
|
13469 |
+
"loss": 0.0046,
|
13470 |
+
"step": 1966
|
13471 |
+
},
|
13472 |
+
{
|
13473 |
+
"epoch": 196.7,
|
13474 |
+
"learning_rate": 9.491052631578948e-05,
|
13475 |
+
"loss": 0.005,
|
13476 |
+
"step": 1967
|
13477 |
+
},
|
13478 |
+
{
|
13479 |
+
"epoch": 196.8,
|
13480 |
+
"learning_rate": 9.490526315789474e-05,
|
13481 |
+
"loss": 0.0054,
|
13482 |
+
"step": 1968
|
13483 |
+
},
|
13484 |
+
{
|
13485 |
+
"epoch": 196.9,
|
13486 |
+
"learning_rate": 9.49e-05,
|
13487 |
+
"loss": 0.0049,
|
13488 |
+
"step": 1969
|
13489 |
+
},
|
13490 |
+
{
|
13491 |
+
"epoch": 197.0,
|
13492 |
+
"learning_rate": 9.489473684210527e-05,
|
13493 |
+
"loss": 0.0057,
|
13494 |
+
"step": 1970
|
13495 |
+
},
|
13496 |
+
{
|
13497 |
+
"epoch": 197.1,
|
13498 |
+
"learning_rate": 9.488947368421053e-05,
|
13499 |
+
"loss": 0.0041,
|
13500 |
+
"step": 1971
|
13501 |
+
},
|
13502 |
+
{
|
13503 |
+
"epoch": 197.2,
|
13504 |
+
"learning_rate": 9.48842105263158e-05,
|
13505 |
+
"loss": 0.0048,
|
13506 |
+
"step": 1972
|
13507 |
+
},
|
13508 |
+
{
|
13509 |
+
"epoch": 197.3,
|
13510 |
+
"learning_rate": 9.487894736842105e-05,
|
13511 |
+
"loss": 0.0053,
|
13512 |
+
"step": 1973
|
13513 |
+
},
|
13514 |
+
{
|
13515 |
+
"epoch": 197.4,
|
13516 |
+
"learning_rate": 9.487368421052633e-05,
|
13517 |
+
"loss": 0.0052,
|
13518 |
+
"step": 1974
|
13519 |
+
},
|
13520 |
+
{
|
13521 |
+
"epoch": 197.5,
|
13522 |
+
"learning_rate": 9.486842105263159e-05,
|
13523 |
+
"loss": 0.005,
|
13524 |
+
"step": 1975
|
13525 |
+
},
|
13526 |
+
{
|
13527 |
+
"epoch": 197.6,
|
13528 |
+
"learning_rate": 9.486315789473684e-05,
|
13529 |
+
"loss": 0.0048,
|
13530 |
+
"step": 1976
|
13531 |
+
},
|
13532 |
+
{
|
13533 |
+
"epoch": 197.7,
|
13534 |
+
"learning_rate": 9.485789473684211e-05,
|
13535 |
+
"loss": 0.005,
|
13536 |
+
"step": 1977
|
13537 |
+
},
|
13538 |
+
{
|
13539 |
+
"epoch": 197.8,
|
13540 |
+
"learning_rate": 9.485263157894738e-05,
|
13541 |
+
"loss": 0.0043,
|
13542 |
+
"step": 1978
|
13543 |
+
},
|
13544 |
+
{
|
13545 |
+
"epoch": 197.9,
|
13546 |
+
"learning_rate": 9.484736842105264e-05,
|
13547 |
+
"loss": 0.0053,
|
13548 |
+
"step": 1979
|
13549 |
+
},
|
13550 |
+
{
|
13551 |
+
"epoch": 198.0,
|
13552 |
+
"learning_rate": 9.48421052631579e-05,
|
13553 |
+
"loss": 0.0049,
|
13554 |
+
"step": 1980
|
13555 |
+
},
|
13556 |
+
{
|
13557 |
+
"epoch": 198.0,
|
13558 |
+
"eval_accuracy_safe": 0.8479477719823748,
|
13559 |
+
"eval_accuracy_unlabeled": NaN,
|
13560 |
+
"eval_accuracy_unsafe": 0.9959019870288063,
|
13561 |
+
"eval_iou_safe": 0.7473294917880173,
|
13562 |
+
"eval_iou_unlabeled": NaN,
|
13563 |
+
"eval_iou_unsafe": 0.9913141008814026,
|
13564 |
+
"eval_loss": 0.045305173844099045,
|
13565 |
+
"eval_mean_accuracy": 0.9219248795055905,
|
13566 |
+
"eval_mean_iou": 0.86932179633471,
|
13567 |
+
"eval_overall_accuracy": 0.991531656749213,
|
13568 |
+
"eval_runtime": 10.1956,
|
13569 |
+
"eval_samples_per_second": 6.571,
|
13570 |
+
"eval_steps_per_second": 0.49,
|
13571 |
+
"step": 1980
|
13572 |
+
},
|
13573 |
+
{
|
13574 |
+
"epoch": 198.1,
|
13575 |
+
"learning_rate": 9.483684210526315e-05,
|
13576 |
+
"loss": 0.0052,
|
13577 |
+
"step": 1981
|
13578 |
+
},
|
13579 |
+
{
|
13580 |
+
"epoch": 198.2,
|
13581 |
+
"learning_rate": 9.483157894736843e-05,
|
13582 |
+
"loss": 0.005,
|
13583 |
+
"step": 1982
|
13584 |
+
},
|
13585 |
+
{
|
13586 |
+
"epoch": 198.3,
|
13587 |
+
"learning_rate": 9.482631578947369e-05,
|
13588 |
+
"loss": 0.0051,
|
13589 |
+
"step": 1983
|
13590 |
+
},
|
13591 |
+
{
|
13592 |
+
"epoch": 198.4,
|
13593 |
+
"learning_rate": 9.482105263157895e-05,
|
13594 |
+
"loss": 0.0046,
|
13595 |
+
"step": 1984
|
13596 |
+
},
|
13597 |
+
{
|
13598 |
+
"epoch": 198.5,
|
13599 |
+
"learning_rate": 9.481578947368421e-05,
|
13600 |
+
"loss": 0.0049,
|
13601 |
+
"step": 1985
|
13602 |
+
},
|
13603 |
+
{
|
13604 |
+
"epoch": 198.6,
|
13605 |
+
"learning_rate": 9.481052631578949e-05,
|
13606 |
+
"loss": 0.0044,
|
13607 |
+
"step": 1986
|
13608 |
+
},
|
13609 |
+
{
|
13610 |
+
"epoch": 198.7,
|
13611 |
+
"learning_rate": 9.480526315789474e-05,
|
13612 |
+
"loss": 0.0049,
|
13613 |
+
"step": 1987
|
13614 |
+
},
|
13615 |
+
{
|
13616 |
+
"epoch": 198.8,
|
13617 |
+
"learning_rate": 9.48e-05,
|
13618 |
+
"loss": 0.005,
|
13619 |
+
"step": 1988
|
13620 |
+
},
|
13621 |
+
{
|
13622 |
+
"epoch": 198.9,
|
13623 |
+
"learning_rate": 9.479473684210528e-05,
|
13624 |
+
"loss": 0.0045,
|
13625 |
+
"step": 1989
|
13626 |
+
},
|
13627 |
+
{
|
13628 |
+
"epoch": 199.0,
|
13629 |
+
"learning_rate": 9.478947368421053e-05,
|
13630 |
+
"loss": 0.005,
|
13631 |
+
"step": 1990
|
13632 |
+
},
|
13633 |
+
{
|
13634 |
+
"epoch": 199.1,
|
13635 |
+
"learning_rate": 9.47842105263158e-05,
|
13636 |
+
"loss": 0.0043,
|
13637 |
+
"step": 1991
|
13638 |
+
},
|
13639 |
+
{
|
13640 |
+
"epoch": 199.2,
|
13641 |
+
"learning_rate": 9.477894736842105e-05,
|
13642 |
+
"loss": 0.0039,
|
13643 |
+
"step": 1992
|
13644 |
+
},
|
13645 |
+
{
|
13646 |
+
"epoch": 199.3,
|
13647 |
+
"learning_rate": 9.477368421052633e-05,
|
13648 |
+
"loss": 0.0058,
|
13649 |
+
"step": 1993
|
13650 |
+
},
|
13651 |
+
{
|
13652 |
+
"epoch": 199.4,
|
13653 |
+
"learning_rate": 9.476842105263159e-05,
|
13654 |
+
"loss": 0.0042,
|
13655 |
+
"step": 1994
|
13656 |
+
},
|
13657 |
+
{
|
13658 |
+
"epoch": 199.5,
|
13659 |
+
"learning_rate": 9.476315789473684e-05,
|
13660 |
+
"loss": 0.0053,
|
13661 |
+
"step": 1995
|
13662 |
+
},
|
13663 |
+
{
|
13664 |
+
"epoch": 199.6,
|
13665 |
+
"learning_rate": 9.475789473684211e-05,
|
13666 |
+
"loss": 0.005,
|
13667 |
+
"step": 1996
|
13668 |
+
},
|
13669 |
+
{
|
13670 |
+
"epoch": 199.7,
|
13671 |
+
"learning_rate": 9.475263157894737e-05,
|
13672 |
+
"loss": 0.0059,
|
13673 |
+
"step": 1997
|
13674 |
+
},
|
13675 |
+
{
|
13676 |
+
"epoch": 199.8,
|
13677 |
+
"learning_rate": 9.474736842105264e-05,
|
13678 |
+
"loss": 0.0048,
|
13679 |
+
"step": 1998
|
13680 |
+
},
|
13681 |
+
{
|
13682 |
+
"epoch": 199.9,
|
13683 |
+
"learning_rate": 9.47421052631579e-05,
|
13684 |
+
"loss": 0.0043,
|
13685 |
+
"step": 1999
|
13686 |
+
},
|
13687 |
+
{
|
13688 |
+
"epoch": 200.0,
|
13689 |
+
"learning_rate": 9.473684210526316e-05,
|
13690 |
+
"loss": 0.0059,
|
13691 |
+
"step": 2000
|
13692 |
+
},
|
13693 |
+
{
|
13694 |
+
"epoch": 200.0,
|
13695 |
+
"eval_accuracy_safe": 0.8505769060258056,
|
13696 |
+
"eval_accuracy_unlabeled": NaN,
|
13697 |
+
"eval_accuracy_unsafe": 0.9960863242765584,
|
13698 |
+
"eval_iou_safe": 0.7536694505644652,
|
13699 |
+
"eval_iou_unlabeled": NaN,
|
13700 |
+
"eval_iou_unsafe": 0.9915765734959425,
|
13701 |
+
"eval_loss": 0.04439888522028923,
|
13702 |
+
"eval_mean_accuracy": 0.923331615151182,
|
13703 |
+
"eval_mean_iou": 0.8726230120302039,
|
13704 |
+
"eval_overall_accuracy": 0.9917882093742713,
|
13705 |
+
"eval_runtime": 11.3486,
|
13706 |
+
"eval_samples_per_second": 5.904,
|
13707 |
+
"eval_steps_per_second": 0.441,
|
13708 |
+
"step": 2000
|
13709 |
+
},
|
13710 |
+
{
|
13711 |
+
"epoch": 200.1,
|
13712 |
+
"learning_rate": 9.473157894736842e-05,
|
13713 |
+
"loss": 0.0042,
|
13714 |
+
"step": 2001
|
13715 |
+
},
|
13716 |
+
{
|
13717 |
+
"epoch": 200.2,
|
13718 |
+
"learning_rate": 9.472631578947369e-05,
|
13719 |
+
"loss": 0.0055,
|
13720 |
+
"step": 2002
|
13721 |
+
},
|
13722 |
+
{
|
13723 |
+
"epoch": 200.3,
|
13724 |
+
"learning_rate": 9.472105263157895e-05,
|
13725 |
+
"loss": 0.005,
|
13726 |
+
"step": 2003
|
13727 |
+
},
|
13728 |
+
{
|
13729 |
+
"epoch": 200.4,
|
13730 |
+
"learning_rate": 9.471578947368421e-05,
|
13731 |
+
"loss": 0.0049,
|
13732 |
+
"step": 2004
|
13733 |
+
},
|
13734 |
+
{
|
13735 |
+
"epoch": 200.5,
|
13736 |
+
"learning_rate": 9.471052631578949e-05,
|
13737 |
+
"loss": 0.0045,
|
13738 |
+
"step": 2005
|
13739 |
+
},
|
13740 |
+
{
|
13741 |
+
"epoch": 200.6,
|
13742 |
+
"learning_rate": 9.470526315789474e-05,
|
13743 |
+
"loss": 0.0056,
|
13744 |
+
"step": 2006
|
13745 |
+
},
|
13746 |
+
{
|
13747 |
+
"epoch": 200.7,
|
13748 |
+
"learning_rate": 9.47e-05,
|
13749 |
+
"loss": 0.0043,
|
13750 |
+
"step": 2007
|
13751 |
+
},
|
13752 |
+
{
|
13753 |
+
"epoch": 200.8,
|
13754 |
+
"learning_rate": 9.469473684210527e-05,
|
13755 |
+
"loss": 0.0048,
|
13756 |
+
"step": 2008
|
13757 |
+
},
|
13758 |
+
{
|
13759 |
+
"epoch": 200.9,
|
13760 |
+
"learning_rate": 9.468947368421052e-05,
|
13761 |
+
"loss": 0.0048,
|
13762 |
+
"step": 2009
|
13763 |
+
},
|
13764 |
+
{
|
13765 |
+
"epoch": 201.0,
|
13766 |
+
"learning_rate": 9.46842105263158e-05,
|
13767 |
+
"loss": 0.0041,
|
13768 |
+
"step": 2010
|
13769 |
+
},
|
13770 |
+
{
|
13771 |
+
"epoch": 201.1,
|
13772 |
+
"learning_rate": 9.467894736842106e-05,
|
13773 |
+
"loss": 0.0049,
|
13774 |
+
"step": 2011
|
13775 |
+
},
|
13776 |
+
{
|
13777 |
+
"epoch": 201.2,
|
13778 |
+
"learning_rate": 9.467368421052632e-05,
|
13779 |
+
"loss": 0.0049,
|
13780 |
+
"step": 2012
|
13781 |
+
},
|
13782 |
+
{
|
13783 |
+
"epoch": 201.3,
|
13784 |
+
"learning_rate": 9.466842105263159e-05,
|
13785 |
+
"loss": 0.0038,
|
13786 |
+
"step": 2013
|
13787 |
+
},
|
13788 |
+
{
|
13789 |
+
"epoch": 201.4,
|
13790 |
+
"learning_rate": 9.466315789473685e-05,
|
13791 |
+
"loss": 0.0046,
|
13792 |
+
"step": 2014
|
13793 |
+
},
|
13794 |
+
{
|
13795 |
+
"epoch": 201.5,
|
13796 |
+
"learning_rate": 9.465789473684211e-05,
|
13797 |
+
"loss": 0.0044,
|
13798 |
+
"step": 2015
|
13799 |
+
},
|
13800 |
+
{
|
13801 |
+
"epoch": 201.6,
|
13802 |
+
"learning_rate": 9.465263157894737e-05,
|
13803 |
+
"loss": 0.0041,
|
13804 |
+
"step": 2016
|
13805 |
+
},
|
13806 |
+
{
|
13807 |
+
"epoch": 201.7,
|
13808 |
+
"learning_rate": 9.464736842105263e-05,
|
13809 |
+
"loss": 0.0058,
|
13810 |
+
"step": 2017
|
13811 |
+
},
|
13812 |
+
{
|
13813 |
+
"epoch": 201.8,
|
13814 |
+
"learning_rate": 9.46421052631579e-05,
|
13815 |
+
"loss": 0.005,
|
13816 |
+
"step": 2018
|
13817 |
+
},
|
13818 |
+
{
|
13819 |
+
"epoch": 201.9,
|
13820 |
+
"learning_rate": 9.463684210526316e-05,
|
13821 |
+
"loss": 0.0052,
|
13822 |
+
"step": 2019
|
13823 |
+
},
|
13824 |
+
{
|
13825 |
+
"epoch": 202.0,
|
13826 |
+
"learning_rate": 9.463157894736842e-05,
|
13827 |
+
"loss": 0.005,
|
13828 |
+
"step": 2020
|
13829 |
+
},
|
13830 |
+
{
|
13831 |
+
"epoch": 202.0,
|
13832 |
+
"eval_accuracy_safe": 0.8554862934221533,
|
13833 |
+
"eval_accuracy_unlabeled": NaN,
|
13834 |
+
"eval_accuracy_unsafe": 0.9958080583420935,
|
13835 |
+
"eval_iou_safe": 0.7519284103596254,
|
13836 |
+
"eval_iou_unlabeled": NaN,
|
13837 |
+
"eval_iou_unsafe": 0.9914470479245693,
|
13838 |
+
"eval_loss": 0.04469138756394386,
|
13839 |
+
"eval_mean_accuracy": 0.9256471758821234,
|
13840 |
+
"eval_mean_iou": 0.8716877291420974,
|
13841 |
+
"eval_overall_accuracy": 0.9916631784012069,
|
13842 |
+
"eval_runtime": 10.1291,
|
13843 |
+
"eval_samples_per_second": 6.615,
|
13844 |
+
"eval_steps_per_second": 0.494,
|
13845 |
+
"step": 2020
|
13846 |
+
},
|
13847 |
+
{
|
13848 |
+
"epoch": 202.1,
|
13849 |
+
"learning_rate": 9.462631578947368e-05,
|
13850 |
+
"loss": 0.0042,
|
13851 |
+
"step": 2021
|
13852 |
+
},
|
13853 |
+
{
|
13854 |
+
"epoch": 202.2,
|
13855 |
+
"learning_rate": 9.462105263157896e-05,
|
13856 |
+
"loss": 0.0046,
|
13857 |
+
"step": 2022
|
13858 |
+
},
|
13859 |
+
{
|
13860 |
+
"epoch": 202.3,
|
13861 |
+
"learning_rate": 9.461578947368421e-05,
|
13862 |
+
"loss": 0.0049,
|
13863 |
+
"step": 2023
|
13864 |
+
},
|
13865 |
+
{
|
13866 |
+
"epoch": 202.4,
|
13867 |
+
"learning_rate": 9.461052631578948e-05,
|
13868 |
+
"loss": 0.0052,
|
13869 |
+
"step": 2024
|
13870 |
+
},
|
13871 |
+
{
|
13872 |
+
"epoch": 202.5,
|
13873 |
+
"learning_rate": 9.460526315789475e-05,
|
13874 |
+
"loss": 0.0051,
|
13875 |
+
"step": 2025
|
13876 |
+
},
|
13877 |
+
{
|
13878 |
+
"epoch": 202.6,
|
13879 |
+
"learning_rate": 9.46e-05,
|
13880 |
+
"loss": 0.0046,
|
13881 |
+
"step": 2026
|
13882 |
+
},
|
13883 |
+
{
|
13884 |
+
"epoch": 202.7,
|
13885 |
+
"learning_rate": 9.459473684210527e-05,
|
13886 |
+
"loss": 0.0045,
|
13887 |
+
"step": 2027
|
13888 |
+
},
|
13889 |
+
{
|
13890 |
+
"epoch": 202.8,
|
13891 |
+
"learning_rate": 9.458947368421053e-05,
|
13892 |
+
"loss": 0.0043,
|
13893 |
+
"step": 2028
|
13894 |
+
},
|
13895 |
+
{
|
13896 |
+
"epoch": 202.9,
|
13897 |
+
"learning_rate": 9.45842105263158e-05,
|
13898 |
+
"loss": 0.0045,
|
13899 |
+
"step": 2029
|
13900 |
+
},
|
13901 |
+
{
|
13902 |
+
"epoch": 203.0,
|
13903 |
+
"learning_rate": 9.457894736842106e-05,
|
13904 |
+
"loss": 0.0057,
|
13905 |
+
"step": 2030
|
13906 |
+
},
|
13907 |
+
{
|
13908 |
+
"epoch": 203.1,
|
13909 |
+
"learning_rate": 9.457368421052632e-05,
|
13910 |
+
"loss": 0.0053,
|
13911 |
+
"step": 2031
|
13912 |
+
},
|
13913 |
+
{
|
13914 |
+
"epoch": 203.2,
|
13915 |
+
"learning_rate": 9.456842105263158e-05,
|
13916 |
+
"loss": 0.0045,
|
13917 |
+
"step": 2032
|
13918 |
+
},
|
13919 |
+
{
|
13920 |
+
"epoch": 203.3,
|
13921 |
+
"learning_rate": 9.456315789473685e-05,
|
13922 |
+
"loss": 0.0045,
|
13923 |
+
"step": 2033
|
13924 |
+
},
|
13925 |
+
{
|
13926 |
+
"epoch": 203.4,
|
13927 |
+
"learning_rate": 9.455789473684211e-05,
|
13928 |
+
"loss": 0.0047,
|
13929 |
+
"step": 2034
|
13930 |
+
},
|
13931 |
+
{
|
13932 |
+
"epoch": 203.5,
|
13933 |
+
"learning_rate": 9.455263157894737e-05,
|
13934 |
+
"loss": 0.0053,
|
13935 |
+
"step": 2035
|
13936 |
+
},
|
13937 |
+
{
|
13938 |
+
"epoch": 203.6,
|
13939 |
+
"learning_rate": 9.454736842105265e-05,
|
13940 |
+
"loss": 0.0044,
|
13941 |
+
"step": 2036
|
13942 |
+
},
|
13943 |
+
{
|
13944 |
+
"epoch": 203.7,
|
13945 |
+
"learning_rate": 9.45421052631579e-05,
|
13946 |
+
"loss": 0.0057,
|
13947 |
+
"step": 2037
|
13948 |
+
},
|
13949 |
+
{
|
13950 |
+
"epoch": 203.8,
|
13951 |
+
"learning_rate": 9.453684210526316e-05,
|
13952 |
+
"loss": 0.0049,
|
13953 |
+
"step": 2038
|
13954 |
+
},
|
13955 |
+
{
|
13956 |
+
"epoch": 203.9,
|
13957 |
+
"learning_rate": 9.453157894736843e-05,
|
13958 |
+
"loss": 0.0041,
|
13959 |
+
"step": 2039
|
13960 |
+
},
|
13961 |
+
{
|
13962 |
+
"epoch": 204.0,
|
13963 |
+
"learning_rate": 9.452631578947368e-05,
|
13964 |
+
"loss": 0.005,
|
13965 |
+
"step": 2040
|
13966 |
+
},
|
13967 |
+
{
|
13968 |
+
"epoch": 204.0,
|
13969 |
+
"eval_accuracy_safe": 0.8493991927556177,
|
13970 |
+
"eval_accuracy_unlabeled": NaN,
|
13971 |
+
"eval_accuracy_unsafe": 0.9959986731472963,
|
13972 |
+
"eval_iou_safe": 0.7507103869820307,
|
13973 |
+
"eval_iou_unlabeled": NaN,
|
13974 |
+
"eval_iou_unsafe": 0.9914539398462113,
|
13975 |
+
"eval_loss": 0.04513503238558769,
|
13976 |
+
"eval_mean_accuracy": 0.922698932951457,
|
13977 |
+
"eval_mean_iou": 0.871082163414121,
|
13978 |
+
"eval_overall_accuracy": 0.9916683595571945,
|
13979 |
+
"eval_runtime": 9.8572,
|
13980 |
+
"eval_samples_per_second": 6.797,
|
13981 |
+
"eval_steps_per_second": 0.507,
|
13982 |
+
"step": 2040
|
13983 |
+
},
|
13984 |
+
{
|
13985 |
+
"epoch": 204.1,
|
13986 |
+
"learning_rate": 9.452105263157896e-05,
|
13987 |
+
"loss": 0.0045,
|
13988 |
+
"step": 2041
|
13989 |
+
},
|
13990 |
+
{
|
13991 |
+
"epoch": 204.2,
|
13992 |
+
"learning_rate": 9.451578947368422e-05,
|
13993 |
+
"loss": 0.0046,
|
13994 |
+
"step": 2042
|
13995 |
+
},
|
13996 |
+
{
|
13997 |
+
"epoch": 204.3,
|
13998 |
+
"learning_rate": 9.451052631578948e-05,
|
13999 |
+
"loss": 0.005,
|
14000 |
+
"step": 2043
|
14001 |
+
},
|
14002 |
+
{
|
14003 |
+
"epoch": 204.4,
|
14004 |
+
"learning_rate": 9.450526315789474e-05,
|
14005 |
+
"loss": 0.0057,
|
14006 |
+
"step": 2044
|
14007 |
+
},
|
14008 |
+
{
|
14009 |
+
"epoch": 204.5,
|
14010 |
+
"learning_rate": 9.449999999999999e-05,
|
14011 |
+
"loss": 0.0041,
|
14012 |
+
"step": 2045
|
14013 |
+
},
|
14014 |
+
{
|
14015 |
+
"epoch": 204.6,
|
14016 |
+
"learning_rate": 9.449473684210527e-05,
|
14017 |
+
"loss": 0.0051,
|
14018 |
+
"step": 2046
|
14019 |
+
},
|
14020 |
+
{
|
14021 |
+
"epoch": 204.7,
|
14022 |
+
"learning_rate": 9.448947368421053e-05,
|
14023 |
+
"loss": 0.0051,
|
14024 |
+
"step": 2047
|
14025 |
+
},
|
14026 |
+
{
|
14027 |
+
"epoch": 204.8,
|
14028 |
+
"learning_rate": 9.44842105263158e-05,
|
14029 |
+
"loss": 0.0049,
|
14030 |
+
"step": 2048
|
14031 |
+
},
|
14032 |
+
{
|
14033 |
+
"epoch": 204.9,
|
14034 |
+
"learning_rate": 9.447894736842106e-05,
|
14035 |
+
"loss": 0.0043,
|
14036 |
+
"step": 2049
|
14037 |
+
},
|
14038 |
+
{
|
14039 |
+
"epoch": 205.0,
|
14040 |
+
"learning_rate": 9.447368421052633e-05,
|
14041 |
+
"loss": 0.0048,
|
14042 |
+
"step": 2050
|
14043 |
+
},
|
14044 |
+
{
|
14045 |
+
"epoch": 205.1,
|
14046 |
+
"learning_rate": 9.446842105263158e-05,
|
14047 |
+
"loss": 0.0038,
|
14048 |
+
"step": 2051
|
14049 |
+
},
|
14050 |
+
{
|
14051 |
+
"epoch": 205.2,
|
14052 |
+
"learning_rate": 9.446315789473684e-05,
|
14053 |
+
"loss": 0.0044,
|
14054 |
+
"step": 2052
|
14055 |
+
},
|
14056 |
+
{
|
14057 |
+
"epoch": 205.3,
|
14058 |
+
"learning_rate": 9.44578947368421e-05,
|
14059 |
+
"loss": 0.0056,
|
14060 |
+
"step": 2053
|
14061 |
+
},
|
14062 |
+
{
|
14063 |
+
"epoch": 205.4,
|
14064 |
+
"learning_rate": 9.445263157894737e-05,
|
14065 |
+
"loss": 0.0058,
|
14066 |
+
"step": 2054
|
14067 |
+
},
|
14068 |
+
{
|
14069 |
+
"epoch": 205.5,
|
14070 |
+
"learning_rate": 9.444736842105264e-05,
|
14071 |
+
"loss": 0.0046,
|
14072 |
+
"step": 2055
|
14073 |
+
},
|
14074 |
+
{
|
14075 |
+
"epoch": 205.6,
|
14076 |
+
"learning_rate": 9.444210526315789e-05,
|
14077 |
+
"loss": 0.0051,
|
14078 |
+
"step": 2056
|
14079 |
+
},
|
14080 |
+
{
|
14081 |
+
"epoch": 205.7,
|
14082 |
+
"learning_rate": 9.443684210526315e-05,
|
14083 |
+
"loss": 0.0042,
|
14084 |
+
"step": 2057
|
14085 |
+
},
|
14086 |
+
{
|
14087 |
+
"epoch": 205.8,
|
14088 |
+
"learning_rate": 9.443157894736843e-05,
|
14089 |
+
"loss": 0.0045,
|
14090 |
+
"step": 2058
|
14091 |
+
},
|
14092 |
+
{
|
14093 |
+
"epoch": 205.9,
|
14094 |
+
"learning_rate": 9.442631578947368e-05,
|
14095 |
+
"loss": 0.0044,
|
14096 |
+
"step": 2059
|
14097 |
+
},
|
14098 |
+
{
|
14099 |
+
"epoch": 206.0,
|
14100 |
+
"learning_rate": 9.442105263157895e-05,
|
14101 |
+
"loss": 0.0043,
|
14102 |
+
"step": 2060
|
14103 |
+
},
|
14104 |
+
{
|
14105 |
+
"epoch": 206.0,
|
14106 |
+
"eval_accuracy_safe": 0.8479554820528834,
|
14107 |
+
"eval_accuracy_unlabeled": NaN,
|
14108 |
+
"eval_accuracy_unsafe": 0.9960219059767392,
|
14109 |
+
"eval_iou_safe": 0.7499403349425852,
|
14110 |
+
"eval_iou_unlabeled": NaN,
|
14111 |
+
"eval_iou_unsafe": 0.9914336989838719,
|
14112 |
+
"eval_loss": 0.04577173292636871,
|
14113 |
+
"eval_mean_accuracy": 0.9219886940148113,
|
14114 |
+
"eval_mean_iou": 0.8706870169632286,
|
14115 |
+
"eval_overall_accuracy": 0.9916482612268248,
|
14116 |
+
"eval_runtime": 9.6732,
|
14117 |
+
"eval_samples_per_second": 6.926,
|
14118 |
+
"eval_steps_per_second": 0.517,
|
14119 |
+
"step": 2060
|
14120 |
+
},
|
14121 |
+
{
|
14122 |
+
"epoch": 206.1,
|
14123 |
+
"learning_rate": 9.441578947368422e-05,
|
14124 |
+
"loss": 0.0046,
|
14125 |
+
"step": 2061
|
14126 |
+
},
|
14127 |
+
{
|
14128 |
+
"epoch": 206.2,
|
14129 |
+
"learning_rate": 9.441052631578948e-05,
|
14130 |
+
"loss": 0.0048,
|
14131 |
+
"step": 2062
|
14132 |
+
},
|
14133 |
+
{
|
14134 |
+
"epoch": 206.3,
|
14135 |
+
"learning_rate": 9.440526315789474e-05,
|
14136 |
+
"loss": 0.0043,
|
14137 |
+
"step": 2063
|
14138 |
+
},
|
14139 |
+
{
|
14140 |
+
"epoch": 206.4,
|
14141 |
+
"learning_rate": 9.44e-05,
|
14142 |
+
"loss": 0.0044,
|
14143 |
+
"step": 2064
|
14144 |
+
},
|
14145 |
+
{
|
14146 |
+
"epoch": 206.5,
|
14147 |
+
"learning_rate": 9.439473684210527e-05,
|
14148 |
+
"loss": 0.0051,
|
14149 |
+
"step": 2065
|
14150 |
+
},
|
14151 |
+
{
|
14152 |
+
"epoch": 206.6,
|
14153 |
+
"learning_rate": 9.438947368421053e-05,
|
14154 |
+
"loss": 0.005,
|
14155 |
+
"step": 2066
|
14156 |
+
},
|
14157 |
+
{
|
14158 |
+
"epoch": 206.7,
|
14159 |
+
"learning_rate": 9.438421052631579e-05,
|
14160 |
+
"loss": 0.0051,
|
14161 |
+
"step": 2067
|
14162 |
+
},
|
14163 |
+
{
|
14164 |
+
"epoch": 206.8,
|
14165 |
+
"learning_rate": 9.437894736842105e-05,
|
14166 |
+
"loss": 0.0043,
|
14167 |
+
"step": 2068
|
14168 |
+
},
|
14169 |
+
{
|
14170 |
+
"epoch": 206.9,
|
14171 |
+
"learning_rate": 9.437368421052633e-05,
|
14172 |
+
"loss": 0.0048,
|
14173 |
+
"step": 2069
|
14174 |
+
},
|
14175 |
+
{
|
14176 |
+
"epoch": 207.0,
|
14177 |
+
"learning_rate": 9.436842105263158e-05,
|
14178 |
+
"loss": 0.0052,
|
14179 |
+
"step": 2070
|
14180 |
+
},
|
14181 |
+
{
|
14182 |
+
"epoch": 207.1,
|
14183 |
+
"learning_rate": 9.436315789473684e-05,
|
14184 |
+
"loss": 0.0045,
|
14185 |
+
"step": 2071
|
14186 |
+
},
|
14187 |
+
{
|
14188 |
+
"epoch": 207.2,
|
14189 |
+
"learning_rate": 9.435789473684212e-05,
|
14190 |
+
"loss": 0.0044,
|
14191 |
+
"step": 2072
|
14192 |
+
},
|
14193 |
+
{
|
14194 |
+
"epoch": 207.3,
|
14195 |
+
"learning_rate": 9.435263157894737e-05,
|
14196 |
+
"loss": 0.0047,
|
14197 |
+
"step": 2073
|
14198 |
+
},
|
14199 |
+
{
|
14200 |
+
"epoch": 207.4,
|
14201 |
+
"learning_rate": 9.434736842105264e-05,
|
14202 |
+
"loss": 0.005,
|
14203 |
+
"step": 2074
|
14204 |
+
},
|
14205 |
+
{
|
14206 |
+
"epoch": 207.5,
|
14207 |
+
"learning_rate": 9.43421052631579e-05,
|
14208 |
+
"loss": 0.0047,
|
14209 |
+
"step": 2075
|
14210 |
+
},
|
14211 |
+
{
|
14212 |
+
"epoch": 207.6,
|
14213 |
+
"learning_rate": 9.433684210526317e-05,
|
14214 |
+
"loss": 0.0057,
|
14215 |
+
"step": 2076
|
14216 |
+
},
|
14217 |
+
{
|
14218 |
+
"epoch": 207.7,
|
14219 |
+
"learning_rate": 9.433157894736843e-05,
|
14220 |
+
"loss": 0.0045,
|
14221 |
+
"step": 2077
|
14222 |
+
},
|
14223 |
+
{
|
14224 |
+
"epoch": 207.8,
|
14225 |
+
"learning_rate": 9.432631578947369e-05,
|
14226 |
+
"loss": 0.0043,
|
14227 |
+
"step": 2078
|
14228 |
+
},
|
14229 |
+
{
|
14230 |
+
"epoch": 207.9,
|
14231 |
+
"learning_rate": 9.432105263157895e-05,
|
14232 |
+
"loss": 0.0049,
|
14233 |
+
"step": 2079
|
14234 |
+
},
|
14235 |
+
{
|
14236 |
+
"epoch": 208.0,
|
14237 |
+
"learning_rate": 9.431578947368421e-05,
|
14238 |
+
"loss": 0.0043,
|
14239 |
+
"step": 2080
|
14240 |
+
},
|
14241 |
+
{
|
14242 |
+
"epoch": 208.0,
|
14243 |
+
"eval_accuracy_safe": 0.8482195519678027,
|
14244 |
+
"eval_accuracy_unlabeled": NaN,
|
14245 |
+
"eval_accuracy_unsafe": 0.9959190596383212,
|
14246 |
+
"eval_iou_safe": 0.7479387654858157,
|
14247 |
+
"eval_iou_unlabeled": NaN,
|
14248 |
+
"eval_iou_unsafe": 0.9913392577140658,
|
14249 |
+
"eval_loss": 0.04617840424180031,
|
14250 |
+
"eval_mean_accuracy": 0.922069305803062,
|
14251 |
+
"eval_mean_iou": 0.8696390115999408,
|
14252 |
+
"eval_overall_accuracy": 0.9915562530062092,
|
14253 |
+
"eval_runtime": 9.8887,
|
14254 |
+
"eval_samples_per_second": 6.775,
|
14255 |
+
"eval_steps_per_second": 0.506,
|
14256 |
+
"step": 2080
|
14257 |
}
|
14258 |
],
|
14259 |
"max_steps": 20000,
|
14260 |
"num_train_epochs": 2000,
|
14261 |
+
"total_flos": 1.2786289100973933e+19,
|
14262 |
"trial_name": null,
|
14263 |
"trial_params": null
|
14264 |
}
|
{checkpoint-1760 β checkpoint-2080}/training_args.bin
RENAMED
File without changes
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:601825867ffe486a1fe2c2d6445b63bf2334be4c983b99fd941f121a72f4c661
|
3 |
size 338796489
|
runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79c74e984fb98f2ed52e15f2d76a4928f473b98641727950291c7ef0bc804f6d
|
3 |
+
size 412129
|