Training in progress, step 1840
Browse files- {checkpoint-1520 β checkpoint-1840}/config.json +0 -0
- {checkpoint-1520 β checkpoint-1840}/optimizer.pt +1 -1
- {checkpoint-1520 β checkpoint-1840}/pytorch_model.bin +1 -1
- {checkpoint-1520 β checkpoint-1840}/rng_state.pth +1 -1
- {checkpoint-1520 β checkpoint-1840}/scheduler.pt +1 -1
- {checkpoint-1520 β checkpoint-1840}/trainer_state.json +2195 -3
- {checkpoint-1520 β checkpoint-1840}/training_args.bin +0 -0
- pytorch_model.bin +1 -1
- runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0 +2 -2
{checkpoint-1520 β checkpoint-1840}/config.json
RENAMED
File without changes
|
{checkpoint-1520 β checkpoint-1840}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 677456319
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3540e7403a48727261433a1412525b9880528c5f8feac23da4d76a8664acca01
|
3 |
size 677456319
|
{checkpoint-1520 β checkpoint-1840}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42e1e3be6e685353b236ae09442bc140dfb5e35929e2ca289bd8856ee6accc20
|
3 |
size 338796489
|
{checkpoint-1520 β checkpoint-1840}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1e19a70789e6b53b575dc7b3ee053801934032d1dc6e1fe536f373f55fbe82c
|
3 |
size 14575
|
{checkpoint-1520 β checkpoint-1840}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b1fff558dc9395dc36231cf87971714a5524894930a2c18a826beb72f565f1c
|
3 |
size 627
|
{checkpoint-1520 β checkpoint-1840}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -10418,11 +10418,2203 @@
|
|
10418 |
"eval_samples_per_second": 6.829,
|
10419 |
"eval_steps_per_second": 0.51,
|
10420 |
"step": 1520
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10421 |
}
|
10422 |
],
|
10423 |
"max_steps": 20000,
|
10424 |
"num_train_epochs": 2000,
|
10425 |
-
"total_flos":
|
10426 |
"trial_name": null,
|
10427 |
"trial_params": null
|
10428 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.03139420226216316,
|
3 |
"best_model_checkpoint": "/robodata/smodak/Projects/nspl/scripts/terrainseg/training/models/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5/checkpoint-1440",
|
4 |
+
"epoch": 184.0,
|
5 |
+
"global_step": 1840,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
10418 |
"eval_samples_per_second": 6.829,
|
10419 |
"eval_steps_per_second": 0.51,
|
10420 |
"step": 1520
|
10421 |
+
},
|
10422 |
+
{
|
10423 |
+
"epoch": 152.1,
|
10424 |
+
"learning_rate": 9.725789473684212e-05,
|
10425 |
+
"loss": 0.0101,
|
10426 |
+
"step": 1521
|
10427 |
+
},
|
10428 |
+
{
|
10429 |
+
"epoch": 152.2,
|
10430 |
+
"learning_rate": 9.725263157894737e-05,
|
10431 |
+
"loss": 0.0081,
|
10432 |
+
"step": 1522
|
10433 |
+
},
|
10434 |
+
{
|
10435 |
+
"epoch": 152.3,
|
10436 |
+
"learning_rate": 9.724736842105263e-05,
|
10437 |
+
"loss": 0.0082,
|
10438 |
+
"step": 1523
|
10439 |
+
},
|
10440 |
+
{
|
10441 |
+
"epoch": 152.4,
|
10442 |
+
"learning_rate": 9.72421052631579e-05,
|
10443 |
+
"loss": 0.0067,
|
10444 |
+
"step": 1524
|
10445 |
+
},
|
10446 |
+
{
|
10447 |
+
"epoch": 152.5,
|
10448 |
+
"learning_rate": 9.723684210526315e-05,
|
10449 |
+
"loss": 0.0071,
|
10450 |
+
"step": 1525
|
10451 |
+
},
|
10452 |
+
{
|
10453 |
+
"epoch": 152.6,
|
10454 |
+
"learning_rate": 9.723157894736843e-05,
|
10455 |
+
"loss": 0.0066,
|
10456 |
+
"step": 1526
|
10457 |
+
},
|
10458 |
+
{
|
10459 |
+
"epoch": 152.7,
|
10460 |
+
"learning_rate": 9.722631578947369e-05,
|
10461 |
+
"loss": 0.0109,
|
10462 |
+
"step": 1527
|
10463 |
+
},
|
10464 |
+
{
|
10465 |
+
"epoch": 152.8,
|
10466 |
+
"learning_rate": 9.722105263157896e-05,
|
10467 |
+
"loss": 0.0062,
|
10468 |
+
"step": 1528
|
10469 |
+
},
|
10470 |
+
{
|
10471 |
+
"epoch": 152.9,
|
10472 |
+
"learning_rate": 9.721578947368422e-05,
|
10473 |
+
"loss": 0.0073,
|
10474 |
+
"step": 1529
|
10475 |
+
},
|
10476 |
+
{
|
10477 |
+
"epoch": 153.0,
|
10478 |
+
"learning_rate": 9.721052631578948e-05,
|
10479 |
+
"loss": 0.0062,
|
10480 |
+
"step": 1530
|
10481 |
+
},
|
10482 |
+
{
|
10483 |
+
"epoch": 153.1,
|
10484 |
+
"learning_rate": 9.720526315789474e-05,
|
10485 |
+
"loss": 0.0105,
|
10486 |
+
"step": 1531
|
10487 |
+
},
|
10488 |
+
{
|
10489 |
+
"epoch": 153.2,
|
10490 |
+
"learning_rate": 9.72e-05,
|
10491 |
+
"loss": 0.0065,
|
10492 |
+
"step": 1532
|
10493 |
+
},
|
10494 |
+
{
|
10495 |
+
"epoch": 153.3,
|
10496 |
+
"learning_rate": 9.719473684210527e-05,
|
10497 |
+
"loss": 0.0066,
|
10498 |
+
"step": 1533
|
10499 |
+
},
|
10500 |
+
{
|
10501 |
+
"epoch": 153.4,
|
10502 |
+
"learning_rate": 9.718947368421053e-05,
|
10503 |
+
"loss": 0.007,
|
10504 |
+
"step": 1534
|
10505 |
+
},
|
10506 |
+
{
|
10507 |
+
"epoch": 153.5,
|
10508 |
+
"learning_rate": 9.718421052631579e-05,
|
10509 |
+
"loss": 0.0071,
|
10510 |
+
"step": 1535
|
10511 |
+
},
|
10512 |
+
{
|
10513 |
+
"epoch": 153.6,
|
10514 |
+
"learning_rate": 9.717894736842105e-05,
|
10515 |
+
"loss": 0.0077,
|
10516 |
+
"step": 1536
|
10517 |
+
},
|
10518 |
+
{
|
10519 |
+
"epoch": 153.7,
|
10520 |
+
"learning_rate": 9.717368421052632e-05,
|
10521 |
+
"loss": 0.008,
|
10522 |
+
"step": 1537
|
10523 |
+
},
|
10524 |
+
{
|
10525 |
+
"epoch": 153.8,
|
10526 |
+
"learning_rate": 9.716842105263159e-05,
|
10527 |
+
"loss": 0.0062,
|
10528 |
+
"step": 1538
|
10529 |
+
},
|
10530 |
+
{
|
10531 |
+
"epoch": 153.9,
|
10532 |
+
"learning_rate": 9.716315789473684e-05,
|
10533 |
+
"loss": 0.0071,
|
10534 |
+
"step": 1539
|
10535 |
+
},
|
10536 |
+
{
|
10537 |
+
"epoch": 154.0,
|
10538 |
+
"learning_rate": 9.715789473684212e-05,
|
10539 |
+
"loss": 0.0078,
|
10540 |
+
"step": 1540
|
10541 |
+
},
|
10542 |
+
{
|
10543 |
+
"epoch": 154.0,
|
10544 |
+
"eval_accuracy_safe": 0.8423425507226263,
|
10545 |
+
"eval_accuracy_unlabeled": NaN,
|
10546 |
+
"eval_accuracy_unsafe": 0.9957350744031362,
|
10547 |
+
"eval_iou_safe": 0.7388186246084089,
|
10548 |
+
"eval_iou_unlabeled": NaN,
|
10549 |
+
"eval_iou_unsafe": 0.9909796662380751,
|
10550 |
+
"eval_loss": 0.04284331202507019,
|
10551 |
+
"eval_mean_accuracy": 0.9190388125628812,
|
10552 |
+
"eval_mean_iou": 0.864899145423242,
|
10553 |
+
"eval_overall_accuracy": 0.9912041052063899,
|
10554 |
+
"eval_runtime": 9.6038,
|
10555 |
+
"eval_samples_per_second": 6.976,
|
10556 |
+
"eval_steps_per_second": 0.521,
|
10557 |
+
"step": 1540
|
10558 |
+
},
|
10559 |
+
{
|
10560 |
+
"epoch": 154.1,
|
10561 |
+
"learning_rate": 9.715263157894738e-05,
|
10562 |
+
"loss": 0.0068,
|
10563 |
+
"step": 1541
|
10564 |
+
},
|
10565 |
+
{
|
10566 |
+
"epoch": 154.2,
|
10567 |
+
"learning_rate": 9.714736842105263e-05,
|
10568 |
+
"loss": 0.0202,
|
10569 |
+
"step": 1542
|
10570 |
+
},
|
10571 |
+
{
|
10572 |
+
"epoch": 154.3,
|
10573 |
+
"learning_rate": 9.71421052631579e-05,
|
10574 |
+
"loss": 0.0068,
|
10575 |
+
"step": 1543
|
10576 |
+
},
|
10577 |
+
{
|
10578 |
+
"epoch": 154.4,
|
10579 |
+
"learning_rate": 9.713684210526317e-05,
|
10580 |
+
"loss": 0.0064,
|
10581 |
+
"step": 1544
|
10582 |
+
},
|
10583 |
+
{
|
10584 |
+
"epoch": 154.5,
|
10585 |
+
"learning_rate": 9.713157894736843e-05,
|
10586 |
+
"loss": 0.006,
|
10587 |
+
"step": 1545
|
10588 |
+
},
|
10589 |
+
{
|
10590 |
+
"epoch": 154.6,
|
10591 |
+
"learning_rate": 9.712631578947369e-05,
|
10592 |
+
"loss": 0.0079,
|
10593 |
+
"step": 1546
|
10594 |
+
},
|
10595 |
+
{
|
10596 |
+
"epoch": 154.7,
|
10597 |
+
"learning_rate": 9.712105263157895e-05,
|
10598 |
+
"loss": 0.0073,
|
10599 |
+
"step": 1547
|
10600 |
+
},
|
10601 |
+
{
|
10602 |
+
"epoch": 154.8,
|
10603 |
+
"learning_rate": 9.711578947368422e-05,
|
10604 |
+
"loss": 0.0103,
|
10605 |
+
"step": 1548
|
10606 |
+
},
|
10607 |
+
{
|
10608 |
+
"epoch": 154.9,
|
10609 |
+
"learning_rate": 9.711052631578948e-05,
|
10610 |
+
"loss": 0.0059,
|
10611 |
+
"step": 1549
|
10612 |
+
},
|
10613 |
+
{
|
10614 |
+
"epoch": 155.0,
|
10615 |
+
"learning_rate": 9.710526315789474e-05,
|
10616 |
+
"loss": 0.0092,
|
10617 |
+
"step": 1550
|
10618 |
+
},
|
10619 |
+
{
|
10620 |
+
"epoch": 155.1,
|
10621 |
+
"learning_rate": 9.71e-05,
|
10622 |
+
"loss": 0.0084,
|
10623 |
+
"step": 1551
|
10624 |
+
},
|
10625 |
+
{
|
10626 |
+
"epoch": 155.2,
|
10627 |
+
"learning_rate": 9.709473684210528e-05,
|
10628 |
+
"loss": 0.0065,
|
10629 |
+
"step": 1552
|
10630 |
+
},
|
10631 |
+
{
|
10632 |
+
"epoch": 155.3,
|
10633 |
+
"learning_rate": 9.708947368421053e-05,
|
10634 |
+
"loss": 0.0091,
|
10635 |
+
"step": 1553
|
10636 |
+
},
|
10637 |
+
{
|
10638 |
+
"epoch": 155.4,
|
10639 |
+
"learning_rate": 9.708421052631579e-05,
|
10640 |
+
"loss": 0.0079,
|
10641 |
+
"step": 1554
|
10642 |
+
},
|
10643 |
+
{
|
10644 |
+
"epoch": 155.5,
|
10645 |
+
"learning_rate": 9.707894736842106e-05,
|
10646 |
+
"loss": 0.0064,
|
10647 |
+
"step": 1555
|
10648 |
+
},
|
10649 |
+
{
|
10650 |
+
"epoch": 155.6,
|
10651 |
+
"learning_rate": 9.707368421052631e-05,
|
10652 |
+
"loss": 0.0079,
|
10653 |
+
"step": 1556
|
10654 |
+
},
|
10655 |
+
{
|
10656 |
+
"epoch": 155.7,
|
10657 |
+
"learning_rate": 9.706842105263159e-05,
|
10658 |
+
"loss": 0.0079,
|
10659 |
+
"step": 1557
|
10660 |
+
},
|
10661 |
+
{
|
10662 |
+
"epoch": 155.8,
|
10663 |
+
"learning_rate": 9.706315789473684e-05,
|
10664 |
+
"loss": 0.0062,
|
10665 |
+
"step": 1558
|
10666 |
+
},
|
10667 |
+
{
|
10668 |
+
"epoch": 155.9,
|
10669 |
+
"learning_rate": 9.705789473684211e-05,
|
10670 |
+
"loss": 0.0076,
|
10671 |
+
"step": 1559
|
10672 |
+
},
|
10673 |
+
{
|
10674 |
+
"epoch": 156.0,
|
10675 |
+
"learning_rate": 9.705263157894738e-05,
|
10676 |
+
"loss": 0.0056,
|
10677 |
+
"step": 1560
|
10678 |
+
},
|
10679 |
+
{
|
10680 |
+
"epoch": 156.0,
|
10681 |
+
"eval_accuracy_safe": 0.837554596936789,
|
10682 |
+
"eval_accuracy_unlabeled": NaN,
|
10683 |
+
"eval_accuracy_unsafe": 0.9964554094533914,
|
10684 |
+
"eval_iou_safe": 0.7501912057442867,
|
10685 |
+
"eval_iou_unlabeled": NaN,
|
10686 |
+
"eval_iou_unsafe": 0.9915527490811993,
|
10687 |
+
"eval_loss": 0.034032076597213745,
|
10688 |
+
"eval_mean_accuracy": 0.9170050031950903,
|
10689 |
+
"eval_mean_iou": 0.8708719774127429,
|
10690 |
+
"eval_overall_accuracy": 0.9917617342365321,
|
10691 |
+
"eval_runtime": 9.6677,
|
10692 |
+
"eval_samples_per_second": 6.93,
|
10693 |
+
"eval_steps_per_second": 0.517,
|
10694 |
+
"step": 1560
|
10695 |
+
},
|
10696 |
+
{
|
10697 |
+
"epoch": 156.1,
|
10698 |
+
"learning_rate": 9.704736842105263e-05,
|
10699 |
+
"loss": 0.0064,
|
10700 |
+
"step": 1561
|
10701 |
+
},
|
10702 |
+
{
|
10703 |
+
"epoch": 156.2,
|
10704 |
+
"learning_rate": 9.70421052631579e-05,
|
10705 |
+
"loss": 0.0068,
|
10706 |
+
"step": 1562
|
10707 |
+
},
|
10708 |
+
{
|
10709 |
+
"epoch": 156.3,
|
10710 |
+
"learning_rate": 9.703684210526316e-05,
|
10711 |
+
"loss": 0.0057,
|
10712 |
+
"step": 1563
|
10713 |
+
},
|
10714 |
+
{
|
10715 |
+
"epoch": 156.4,
|
10716 |
+
"learning_rate": 9.703157894736843e-05,
|
10717 |
+
"loss": 0.0064,
|
10718 |
+
"step": 1564
|
10719 |
+
},
|
10720 |
+
{
|
10721 |
+
"epoch": 156.5,
|
10722 |
+
"learning_rate": 9.702631578947369e-05,
|
10723 |
+
"loss": 0.008,
|
10724 |
+
"step": 1565
|
10725 |
+
},
|
10726 |
+
{
|
10727 |
+
"epoch": 156.6,
|
10728 |
+
"learning_rate": 9.702105263157895e-05,
|
10729 |
+
"loss": 0.0079,
|
10730 |
+
"step": 1566
|
10731 |
+
},
|
10732 |
+
{
|
10733 |
+
"epoch": 156.7,
|
10734 |
+
"learning_rate": 9.701578947368421e-05,
|
10735 |
+
"loss": 0.0062,
|
10736 |
+
"step": 1567
|
10737 |
+
},
|
10738 |
+
{
|
10739 |
+
"epoch": 156.8,
|
10740 |
+
"learning_rate": 9.701052631578948e-05,
|
10741 |
+
"loss": 0.008,
|
10742 |
+
"step": 1568
|
10743 |
+
},
|
10744 |
+
{
|
10745 |
+
"epoch": 156.9,
|
10746 |
+
"learning_rate": 9.700526315789474e-05,
|
10747 |
+
"loss": 0.0072,
|
10748 |
+
"step": 1569
|
10749 |
+
},
|
10750 |
+
{
|
10751 |
+
"epoch": 157.0,
|
10752 |
+
"learning_rate": 9.7e-05,
|
10753 |
+
"loss": 0.0093,
|
10754 |
+
"step": 1570
|
10755 |
+
},
|
10756 |
+
{
|
10757 |
+
"epoch": 157.1,
|
10758 |
+
"learning_rate": 9.699473684210528e-05,
|
10759 |
+
"loss": 0.0064,
|
10760 |
+
"step": 1571
|
10761 |
+
},
|
10762 |
+
{
|
10763 |
+
"epoch": 157.2,
|
10764 |
+
"learning_rate": 9.698947368421052e-05,
|
10765 |
+
"loss": 0.0059,
|
10766 |
+
"step": 1572
|
10767 |
+
},
|
10768 |
+
{
|
10769 |
+
"epoch": 157.3,
|
10770 |
+
"learning_rate": 9.69842105263158e-05,
|
10771 |
+
"loss": 0.0065,
|
10772 |
+
"step": 1573
|
10773 |
+
},
|
10774 |
+
{
|
10775 |
+
"epoch": 157.4,
|
10776 |
+
"learning_rate": 9.697894736842106e-05,
|
10777 |
+
"loss": 0.0069,
|
10778 |
+
"step": 1574
|
10779 |
+
},
|
10780 |
+
{
|
10781 |
+
"epoch": 157.5,
|
10782 |
+
"learning_rate": 9.697368421052631e-05,
|
10783 |
+
"loss": 0.0084,
|
10784 |
+
"step": 1575
|
10785 |
+
},
|
10786 |
+
{
|
10787 |
+
"epoch": 157.6,
|
10788 |
+
"learning_rate": 9.696842105263159e-05,
|
10789 |
+
"loss": 0.0055,
|
10790 |
+
"step": 1576
|
10791 |
+
},
|
10792 |
+
{
|
10793 |
+
"epoch": 157.7,
|
10794 |
+
"learning_rate": 9.696315789473685e-05,
|
10795 |
+
"loss": 0.0077,
|
10796 |
+
"step": 1577
|
10797 |
+
},
|
10798 |
+
{
|
10799 |
+
"epoch": 157.8,
|
10800 |
+
"learning_rate": 9.695789473684211e-05,
|
10801 |
+
"loss": 0.0081,
|
10802 |
+
"step": 1578
|
10803 |
+
},
|
10804 |
+
{
|
10805 |
+
"epoch": 157.9,
|
10806 |
+
"learning_rate": 9.695263157894737e-05,
|
10807 |
+
"loss": 0.0064,
|
10808 |
+
"step": 1579
|
10809 |
+
},
|
10810 |
+
{
|
10811 |
+
"epoch": 158.0,
|
10812 |
+
"learning_rate": 9.694736842105264e-05,
|
10813 |
+
"loss": 0.0063,
|
10814 |
+
"step": 1580
|
10815 |
+
},
|
10816 |
+
{
|
10817 |
+
"epoch": 158.0,
|
10818 |
+
"eval_accuracy_safe": 0.8444801677711342,
|
10819 |
+
"eval_accuracy_unlabeled": NaN,
|
10820 |
+
"eval_accuracy_unsafe": 0.9957573098636385,
|
10821 |
+
"eval_iou_safe": 0.7411684299919813,
|
10822 |
+
"eval_iou_unlabeled": NaN,
|
10823 |
+
"eval_iou_unsafe": 0.9910659699291059,
|
10824 |
+
"eval_loss": 0.03585521876811981,
|
10825 |
+
"eval_mean_accuracy": 0.9201187388173864,
|
10826 |
+
"eval_mean_iou": 0.8661171999605436,
|
10827 |
+
"eval_overall_accuracy": 0.9912888256471548,
|
10828 |
+
"eval_runtime": 9.6749,
|
10829 |
+
"eval_samples_per_second": 6.925,
|
10830 |
+
"eval_steps_per_second": 0.517,
|
10831 |
+
"step": 1580
|
10832 |
+
},
|
10833 |
+
{
|
10834 |
+
"epoch": 158.1,
|
10835 |
+
"learning_rate": 9.69421052631579e-05,
|
10836 |
+
"loss": 0.0068,
|
10837 |
+
"step": 1581
|
10838 |
+
},
|
10839 |
+
{
|
10840 |
+
"epoch": 158.2,
|
10841 |
+
"learning_rate": 9.693684210526316e-05,
|
10842 |
+
"loss": 0.0048,
|
10843 |
+
"step": 1582
|
10844 |
+
},
|
10845 |
+
{
|
10846 |
+
"epoch": 158.3,
|
10847 |
+
"learning_rate": 9.693157894736842e-05,
|
10848 |
+
"loss": 0.0052,
|
10849 |
+
"step": 1583
|
10850 |
+
},
|
10851 |
+
{
|
10852 |
+
"epoch": 158.4,
|
10853 |
+
"learning_rate": 9.692631578947369e-05,
|
10854 |
+
"loss": 0.0069,
|
10855 |
+
"step": 1584
|
10856 |
+
},
|
10857 |
+
{
|
10858 |
+
"epoch": 158.5,
|
10859 |
+
"learning_rate": 9.692105263157896e-05,
|
10860 |
+
"loss": 0.0073,
|
10861 |
+
"step": 1585
|
10862 |
+
},
|
10863 |
+
{
|
10864 |
+
"epoch": 158.6,
|
10865 |
+
"learning_rate": 9.691578947368421e-05,
|
10866 |
+
"loss": 0.0076,
|
10867 |
+
"step": 1586
|
10868 |
+
},
|
10869 |
+
{
|
10870 |
+
"epoch": 158.7,
|
10871 |
+
"learning_rate": 9.691052631578947e-05,
|
10872 |
+
"loss": 0.0065,
|
10873 |
+
"step": 1587
|
10874 |
+
},
|
10875 |
+
{
|
10876 |
+
"epoch": 158.8,
|
10877 |
+
"learning_rate": 9.690526315789475e-05,
|
10878 |
+
"loss": 0.0069,
|
10879 |
+
"step": 1588
|
10880 |
+
},
|
10881 |
+
{
|
10882 |
+
"epoch": 158.9,
|
10883 |
+
"learning_rate": 9.69e-05,
|
10884 |
+
"loss": 0.0064,
|
10885 |
+
"step": 1589
|
10886 |
+
},
|
10887 |
+
{
|
10888 |
+
"epoch": 159.0,
|
10889 |
+
"learning_rate": 9.689473684210527e-05,
|
10890 |
+
"loss": 0.0073,
|
10891 |
+
"step": 1590
|
10892 |
+
},
|
10893 |
+
{
|
10894 |
+
"epoch": 159.1,
|
10895 |
+
"learning_rate": 9.688947368421054e-05,
|
10896 |
+
"loss": 0.0058,
|
10897 |
+
"step": 1591
|
10898 |
+
},
|
10899 |
+
{
|
10900 |
+
"epoch": 159.2,
|
10901 |
+
"learning_rate": 9.68842105263158e-05,
|
10902 |
+
"loss": 0.0067,
|
10903 |
+
"step": 1592
|
10904 |
+
},
|
10905 |
+
{
|
10906 |
+
"epoch": 159.3,
|
10907 |
+
"learning_rate": 9.687894736842106e-05,
|
10908 |
+
"loss": 0.0061,
|
10909 |
+
"step": 1593
|
10910 |
+
},
|
10911 |
+
{
|
10912 |
+
"epoch": 159.4,
|
10913 |
+
"learning_rate": 9.687368421052632e-05,
|
10914 |
+
"loss": 0.0058,
|
10915 |
+
"step": 1594
|
10916 |
+
},
|
10917 |
+
{
|
10918 |
+
"epoch": 159.5,
|
10919 |
+
"learning_rate": 9.686842105263158e-05,
|
10920 |
+
"loss": 0.006,
|
10921 |
+
"step": 1595
|
10922 |
+
},
|
10923 |
+
{
|
10924 |
+
"epoch": 159.6,
|
10925 |
+
"learning_rate": 9.686315789473685e-05,
|
10926 |
+
"loss": 0.006,
|
10927 |
+
"step": 1596
|
10928 |
+
},
|
10929 |
+
{
|
10930 |
+
"epoch": 159.7,
|
10931 |
+
"learning_rate": 9.685789473684211e-05,
|
10932 |
+
"loss": 0.0053,
|
10933 |
+
"step": 1597
|
10934 |
+
},
|
10935 |
+
{
|
10936 |
+
"epoch": 159.8,
|
10937 |
+
"learning_rate": 9.685263157894737e-05,
|
10938 |
+
"loss": 0.0064,
|
10939 |
+
"step": 1598
|
10940 |
+
},
|
10941 |
+
{
|
10942 |
+
"epoch": 159.9,
|
10943 |
+
"learning_rate": 9.684736842105263e-05,
|
10944 |
+
"loss": 0.0077,
|
10945 |
+
"step": 1599
|
10946 |
+
},
|
10947 |
+
{
|
10948 |
+
"epoch": 160.0,
|
10949 |
+
"learning_rate": 9.68421052631579e-05,
|
10950 |
+
"loss": 0.0083,
|
10951 |
+
"step": 1600
|
10952 |
+
},
|
10953 |
+
{
|
10954 |
+
"epoch": 160.0,
|
10955 |
+
"eval_accuracy_safe": 0.8410125635598937,
|
10956 |
+
"eval_accuracy_unlabeled": NaN,
|
10957 |
+
"eval_accuracy_unsafe": 0.9961029862047448,
|
10958 |
+
"eval_iou_safe": 0.7455564175207526,
|
10959 |
+
"eval_iou_unlabeled": NaN,
|
10960 |
+
"eval_iou_unsafe": 0.9913058831298214,
|
10961 |
+
"eval_loss": 0.037494219839572906,
|
10962 |
+
"eval_mean_accuracy": 0.9185577748823193,
|
10963 |
+
"eval_mean_iou": 0.8684311503252871,
|
10964 |
+
"eval_overall_accuracy": 0.9915218637950385,
|
10965 |
+
"eval_runtime": 9.7803,
|
10966 |
+
"eval_samples_per_second": 6.851,
|
10967 |
+
"eval_steps_per_second": 0.511,
|
10968 |
+
"step": 1600
|
10969 |
+
},
|
10970 |
+
{
|
10971 |
+
"epoch": 160.1,
|
10972 |
+
"learning_rate": 9.683684210526316e-05,
|
10973 |
+
"loss": 0.0068,
|
10974 |
+
"step": 1601
|
10975 |
+
},
|
10976 |
+
{
|
10977 |
+
"epoch": 160.2,
|
10978 |
+
"learning_rate": 9.683157894736843e-05,
|
10979 |
+
"loss": 0.0062,
|
10980 |
+
"step": 1602
|
10981 |
+
},
|
10982 |
+
{
|
10983 |
+
"epoch": 160.3,
|
10984 |
+
"learning_rate": 9.682631578947368e-05,
|
10985 |
+
"loss": 0.0068,
|
10986 |
+
"step": 1603
|
10987 |
+
},
|
10988 |
+
{
|
10989 |
+
"epoch": 160.4,
|
10990 |
+
"learning_rate": 9.682105263157896e-05,
|
10991 |
+
"loss": 0.0062,
|
10992 |
+
"step": 1604
|
10993 |
+
},
|
10994 |
+
{
|
10995 |
+
"epoch": 160.5,
|
10996 |
+
"learning_rate": 9.681578947368422e-05,
|
10997 |
+
"loss": 0.0056,
|
10998 |
+
"step": 1605
|
10999 |
+
},
|
11000 |
+
{
|
11001 |
+
"epoch": 160.6,
|
11002 |
+
"learning_rate": 9.681052631578947e-05,
|
11003 |
+
"loss": 0.0059,
|
11004 |
+
"step": 1606
|
11005 |
+
},
|
11006 |
+
{
|
11007 |
+
"epoch": 160.7,
|
11008 |
+
"learning_rate": 9.680526315789475e-05,
|
11009 |
+
"loss": 0.0056,
|
11010 |
+
"step": 1607
|
11011 |
+
},
|
11012 |
+
{
|
11013 |
+
"epoch": 160.8,
|
11014 |
+
"learning_rate": 9.680000000000001e-05,
|
11015 |
+
"loss": 0.0073,
|
11016 |
+
"step": 1608
|
11017 |
+
},
|
11018 |
+
{
|
11019 |
+
"epoch": 160.9,
|
11020 |
+
"learning_rate": 9.679473684210527e-05,
|
11021 |
+
"loss": 0.0069,
|
11022 |
+
"step": 1609
|
11023 |
+
},
|
11024 |
+
{
|
11025 |
+
"epoch": 161.0,
|
11026 |
+
"learning_rate": 9.678947368421053e-05,
|
11027 |
+
"loss": 0.0062,
|
11028 |
+
"step": 1610
|
11029 |
+
},
|
11030 |
+
{
|
11031 |
+
"epoch": 161.1,
|
11032 |
+
"learning_rate": 9.67842105263158e-05,
|
11033 |
+
"loss": 0.0082,
|
11034 |
+
"step": 1611
|
11035 |
+
},
|
11036 |
+
{
|
11037 |
+
"epoch": 161.2,
|
11038 |
+
"learning_rate": 9.677894736842106e-05,
|
11039 |
+
"loss": 0.0059,
|
11040 |
+
"step": 1612
|
11041 |
+
},
|
11042 |
+
{
|
11043 |
+
"epoch": 161.3,
|
11044 |
+
"learning_rate": 9.677368421052632e-05,
|
11045 |
+
"loss": 0.0055,
|
11046 |
+
"step": 1613
|
11047 |
+
},
|
11048 |
+
{
|
11049 |
+
"epoch": 161.4,
|
11050 |
+
"learning_rate": 9.676842105263158e-05,
|
11051 |
+
"loss": 0.0065,
|
11052 |
+
"step": 1614
|
11053 |
+
},
|
11054 |
+
{
|
11055 |
+
"epoch": 161.5,
|
11056 |
+
"learning_rate": 9.676315789473684e-05,
|
11057 |
+
"loss": 0.0057,
|
11058 |
+
"step": 1615
|
11059 |
+
},
|
11060 |
+
{
|
11061 |
+
"epoch": 161.6,
|
11062 |
+
"learning_rate": 9.675789473684212e-05,
|
11063 |
+
"loss": 0.0061,
|
11064 |
+
"step": 1616
|
11065 |
+
},
|
11066 |
+
{
|
11067 |
+
"epoch": 161.7,
|
11068 |
+
"learning_rate": 9.675263157894737e-05,
|
11069 |
+
"loss": 0.0061,
|
11070 |
+
"step": 1617
|
11071 |
+
},
|
11072 |
+
{
|
11073 |
+
"epoch": 161.8,
|
11074 |
+
"learning_rate": 9.674736842105263e-05,
|
11075 |
+
"loss": 0.0056,
|
11076 |
+
"step": 1618
|
11077 |
+
},
|
11078 |
+
{
|
11079 |
+
"epoch": 161.9,
|
11080 |
+
"learning_rate": 9.67421052631579e-05,
|
11081 |
+
"loss": 0.0053,
|
11082 |
+
"step": 1619
|
11083 |
+
},
|
11084 |
+
{
|
11085 |
+
"epoch": 162.0,
|
11086 |
+
"learning_rate": 9.673684210526316e-05,
|
11087 |
+
"loss": 0.0065,
|
11088 |
+
"step": 1620
|
11089 |
+
},
|
11090 |
+
{
|
11091 |
+
"epoch": 162.0,
|
11092 |
+
"eval_accuracy_safe": 0.84590267577997,
|
11093 |
+
"eval_accuracy_unlabeled": NaN,
|
11094 |
+
"eval_accuracy_unsafe": 0.9960353411230586,
|
11095 |
+
"eval_iou_safe": 0.7484169794620885,
|
11096 |
+
"eval_iou_unlabeled": NaN,
|
11097 |
+
"eval_iou_unsafe": 0.9913854136155347,
|
11098 |
+
"eval_loss": 0.03703925758600235,
|
11099 |
+
"eval_mean_accuracy": 0.9209690084515143,
|
11100 |
+
"eval_mean_iou": 0.8699011965388116,
|
11101 |
+
"eval_overall_accuracy": 0.9916006629146746,
|
11102 |
+
"eval_runtime": 9.6157,
|
11103 |
+
"eval_samples_per_second": 6.968,
|
11104 |
+
"eval_steps_per_second": 0.52,
|
11105 |
+
"step": 1620
|
11106 |
+
},
|
11107 |
+
{
|
11108 |
+
"epoch": 162.1,
|
11109 |
+
"learning_rate": 9.673157894736843e-05,
|
11110 |
+
"loss": 0.0056,
|
11111 |
+
"step": 1621
|
11112 |
+
},
|
11113 |
+
{
|
11114 |
+
"epoch": 162.2,
|
11115 |
+
"learning_rate": 9.672631578947368e-05,
|
11116 |
+
"loss": 0.0072,
|
11117 |
+
"step": 1622
|
11118 |
+
},
|
11119 |
+
{
|
11120 |
+
"epoch": 162.3,
|
11121 |
+
"learning_rate": 9.672105263157896e-05,
|
11122 |
+
"loss": 0.0064,
|
11123 |
+
"step": 1623
|
11124 |
+
},
|
11125 |
+
{
|
11126 |
+
"epoch": 162.4,
|
11127 |
+
"learning_rate": 9.671578947368422e-05,
|
11128 |
+
"loss": 0.0056,
|
11129 |
+
"step": 1624
|
11130 |
+
},
|
11131 |
+
{
|
11132 |
+
"epoch": 162.5,
|
11133 |
+
"learning_rate": 9.671052631578947e-05,
|
11134 |
+
"loss": 0.0057,
|
11135 |
+
"step": 1625
|
11136 |
+
},
|
11137 |
+
{
|
11138 |
+
"epoch": 162.6,
|
11139 |
+
"learning_rate": 9.670526315789474e-05,
|
11140 |
+
"loss": 0.0062,
|
11141 |
+
"step": 1626
|
11142 |
+
},
|
11143 |
+
{
|
11144 |
+
"epoch": 162.7,
|
11145 |
+
"learning_rate": 9.67e-05,
|
11146 |
+
"loss": 0.0049,
|
11147 |
+
"step": 1627
|
11148 |
+
},
|
11149 |
+
{
|
11150 |
+
"epoch": 162.8,
|
11151 |
+
"learning_rate": 9.669473684210527e-05,
|
11152 |
+
"loss": 0.0058,
|
11153 |
+
"step": 1628
|
11154 |
+
},
|
11155 |
+
{
|
11156 |
+
"epoch": 162.9,
|
11157 |
+
"learning_rate": 9.668947368421053e-05,
|
11158 |
+
"loss": 0.0075,
|
11159 |
+
"step": 1629
|
11160 |
+
},
|
11161 |
+
{
|
11162 |
+
"epoch": 163.0,
|
11163 |
+
"learning_rate": 9.668421052631579e-05,
|
11164 |
+
"loss": 0.0059,
|
11165 |
+
"step": 1630
|
11166 |
+
},
|
11167 |
+
{
|
11168 |
+
"epoch": 163.1,
|
11169 |
+
"learning_rate": 9.667894736842106e-05,
|
11170 |
+
"loss": 0.0074,
|
11171 |
+
"step": 1631
|
11172 |
+
},
|
11173 |
+
{
|
11174 |
+
"epoch": 163.2,
|
11175 |
+
"learning_rate": 9.667368421052632e-05,
|
11176 |
+
"loss": 0.0068,
|
11177 |
+
"step": 1632
|
11178 |
+
},
|
11179 |
+
{
|
11180 |
+
"epoch": 163.3,
|
11181 |
+
"learning_rate": 9.666842105263158e-05,
|
11182 |
+
"loss": 0.0052,
|
11183 |
+
"step": 1633
|
11184 |
+
},
|
11185 |
+
{
|
11186 |
+
"epoch": 163.4,
|
11187 |
+
"learning_rate": 9.666315789473684e-05,
|
11188 |
+
"loss": 0.0052,
|
11189 |
+
"step": 1634
|
11190 |
+
},
|
11191 |
+
{
|
11192 |
+
"epoch": 163.5,
|
11193 |
+
"learning_rate": 9.665789473684212e-05,
|
11194 |
+
"loss": 0.0067,
|
11195 |
+
"step": 1635
|
11196 |
+
},
|
11197 |
+
{
|
11198 |
+
"epoch": 163.6,
|
11199 |
+
"learning_rate": 9.665263157894737e-05,
|
11200 |
+
"loss": 0.0069,
|
11201 |
+
"step": 1636
|
11202 |
+
},
|
11203 |
+
{
|
11204 |
+
"epoch": 163.7,
|
11205 |
+
"learning_rate": 9.664736842105263e-05,
|
11206 |
+
"loss": 0.0053,
|
11207 |
+
"step": 1637
|
11208 |
+
},
|
11209 |
+
{
|
11210 |
+
"epoch": 163.8,
|
11211 |
+
"learning_rate": 9.66421052631579e-05,
|
11212 |
+
"loss": 0.0055,
|
11213 |
+
"step": 1638
|
11214 |
+
},
|
11215 |
+
{
|
11216 |
+
"epoch": 163.9,
|
11217 |
+
"learning_rate": 9.663684210526315e-05,
|
11218 |
+
"loss": 0.0045,
|
11219 |
+
"step": 1639
|
11220 |
+
},
|
11221 |
+
{
|
11222 |
+
"epoch": 164.0,
|
11223 |
+
"learning_rate": 9.663157894736843e-05,
|
11224 |
+
"loss": 0.0063,
|
11225 |
+
"step": 1640
|
11226 |
+
},
|
11227 |
+
{
|
11228 |
+
"epoch": 164.0,
|
11229 |
+
"eval_accuracy_safe": 0.8497769862105389,
|
11230 |
+
"eval_accuracy_unlabeled": NaN,
|
11231 |
+
"eval_accuracy_unsafe": 0.9958769941365266,
|
11232 |
+
"eval_iou_safe": 0.748400042099688,
|
11233 |
+
"eval_iou_unlabeled": NaN,
|
11234 |
+
"eval_iou_unsafe": 0.9913441635827636,
|
11235 |
+
"eval_loss": 0.038791533559560776,
|
11236 |
+
"eval_mean_accuracy": 0.9228269901735328,
|
11237 |
+
"eval_mean_iou": 0.8698721028412257,
|
11238 |
+
"eval_overall_accuracy": 0.9915614341621968,
|
11239 |
+
"eval_runtime": 9.8481,
|
11240 |
+
"eval_samples_per_second": 6.803,
|
11241 |
+
"eval_steps_per_second": 0.508,
|
11242 |
+
"step": 1640
|
11243 |
+
},
|
11244 |
+
{
|
11245 |
+
"epoch": 164.1,
|
11246 |
+
"learning_rate": 9.662631578947369e-05,
|
11247 |
+
"loss": 0.0071,
|
11248 |
+
"step": 1641
|
11249 |
+
},
|
11250 |
+
{
|
11251 |
+
"epoch": 164.2,
|
11252 |
+
"learning_rate": 9.662105263157895e-05,
|
11253 |
+
"loss": 0.006,
|
11254 |
+
"step": 1642
|
11255 |
+
},
|
11256 |
+
{
|
11257 |
+
"epoch": 164.3,
|
11258 |
+
"learning_rate": 9.661578947368422e-05,
|
11259 |
+
"loss": 0.0062,
|
11260 |
+
"step": 1643
|
11261 |
+
},
|
11262 |
+
{
|
11263 |
+
"epoch": 164.4,
|
11264 |
+
"learning_rate": 9.661052631578948e-05,
|
11265 |
+
"loss": 0.0056,
|
11266 |
+
"step": 1644
|
11267 |
+
},
|
11268 |
+
{
|
11269 |
+
"epoch": 164.5,
|
11270 |
+
"learning_rate": 9.660526315789474e-05,
|
11271 |
+
"loss": 0.0057,
|
11272 |
+
"step": 1645
|
11273 |
+
},
|
11274 |
+
{
|
11275 |
+
"epoch": 164.6,
|
11276 |
+
"learning_rate": 9.66e-05,
|
11277 |
+
"loss": 0.0053,
|
11278 |
+
"step": 1646
|
11279 |
+
},
|
11280 |
+
{
|
11281 |
+
"epoch": 164.7,
|
11282 |
+
"learning_rate": 9.659473684210527e-05,
|
11283 |
+
"loss": 0.0062,
|
11284 |
+
"step": 1647
|
11285 |
+
},
|
11286 |
+
{
|
11287 |
+
"epoch": 164.8,
|
11288 |
+
"learning_rate": 9.658947368421053e-05,
|
11289 |
+
"loss": 0.0062,
|
11290 |
+
"step": 1648
|
11291 |
+
},
|
11292 |
+
{
|
11293 |
+
"epoch": 164.9,
|
11294 |
+
"learning_rate": 9.65842105263158e-05,
|
11295 |
+
"loss": 0.0055,
|
11296 |
+
"step": 1649
|
11297 |
+
},
|
11298 |
+
{
|
11299 |
+
"epoch": 165.0,
|
11300 |
+
"learning_rate": 9.657894736842105e-05,
|
11301 |
+
"loss": 0.0059,
|
11302 |
+
"step": 1650
|
11303 |
+
},
|
11304 |
+
{
|
11305 |
+
"epoch": 165.1,
|
11306 |
+
"learning_rate": 9.657368421052632e-05,
|
11307 |
+
"loss": 0.0059,
|
11308 |
+
"step": 1651
|
11309 |
+
},
|
11310 |
+
{
|
11311 |
+
"epoch": 165.2,
|
11312 |
+
"learning_rate": 9.656842105263159e-05,
|
11313 |
+
"loss": 0.0061,
|
11314 |
+
"step": 1652
|
11315 |
+
},
|
11316 |
+
{
|
11317 |
+
"epoch": 165.3,
|
11318 |
+
"learning_rate": 9.656315789473684e-05,
|
11319 |
+
"loss": 0.0057,
|
11320 |
+
"step": 1653
|
11321 |
+
},
|
11322 |
+
{
|
11323 |
+
"epoch": 165.4,
|
11324 |
+
"learning_rate": 9.655789473684212e-05,
|
11325 |
+
"loss": 0.0066,
|
11326 |
+
"step": 1654
|
11327 |
+
},
|
11328 |
+
{
|
11329 |
+
"epoch": 165.5,
|
11330 |
+
"learning_rate": 9.655263157894738e-05,
|
11331 |
+
"loss": 0.0064,
|
11332 |
+
"step": 1655
|
11333 |
+
},
|
11334 |
+
{
|
11335 |
+
"epoch": 165.6,
|
11336 |
+
"learning_rate": 9.654736842105263e-05,
|
11337 |
+
"loss": 0.0064,
|
11338 |
+
"step": 1656
|
11339 |
+
},
|
11340 |
+
{
|
11341 |
+
"epoch": 165.7,
|
11342 |
+
"learning_rate": 9.65421052631579e-05,
|
11343 |
+
"loss": 0.0058,
|
11344 |
+
"step": 1657
|
11345 |
+
},
|
11346 |
+
{
|
11347 |
+
"epoch": 165.8,
|
11348 |
+
"learning_rate": 9.653684210526316e-05,
|
11349 |
+
"loss": 0.0051,
|
11350 |
+
"step": 1658
|
11351 |
+
},
|
11352 |
+
{
|
11353 |
+
"epoch": 165.9,
|
11354 |
+
"learning_rate": 9.653157894736843e-05,
|
11355 |
+
"loss": 0.0068,
|
11356 |
+
"step": 1659
|
11357 |
+
},
|
11358 |
+
{
|
11359 |
+
"epoch": 166.0,
|
11360 |
+
"learning_rate": 9.652631578947369e-05,
|
11361 |
+
"loss": 0.0056,
|
11362 |
+
"step": 1660
|
11363 |
+
},
|
11364 |
+
{
|
11365 |
+
"epoch": 166.0,
|
11366 |
+
"eval_accuracy_safe": 0.851735344119722,
|
11367 |
+
"eval_accuracy_unlabeled": NaN,
|
11368 |
+
"eval_accuracy_unsafe": 0.9958300004587897,
|
11369 |
+
"eval_iou_safe": 0.7491061726218428,
|
11370 |
+
"eval_iou_unlabeled": NaN,
|
11371 |
+
"eval_iou_unsafe": 0.9913562070641545,
|
11372 |
+
"eval_loss": 0.038572605699300766,
|
11373 |
+
"eval_mean_accuracy": 0.9237826722892558,
|
11374 |
+
"eval_mean_iou": 0.8702311898429986,
|
11375 |
+
"eval_overall_accuracy": 0.9915736753549149,
|
11376 |
+
"eval_runtime": 9.7779,
|
11377 |
+
"eval_samples_per_second": 6.852,
|
11378 |
+
"eval_steps_per_second": 0.511,
|
11379 |
+
"step": 1660
|
11380 |
+
},
|
11381 |
+
{
|
11382 |
+
"epoch": 166.1,
|
11383 |
+
"learning_rate": 9.652105263157895e-05,
|
11384 |
+
"loss": 0.0065,
|
11385 |
+
"step": 1661
|
11386 |
+
},
|
11387 |
+
{
|
11388 |
+
"epoch": 166.2,
|
11389 |
+
"learning_rate": 9.651578947368421e-05,
|
11390 |
+
"loss": 0.0065,
|
11391 |
+
"step": 1662
|
11392 |
+
},
|
11393 |
+
{
|
11394 |
+
"epoch": 166.3,
|
11395 |
+
"learning_rate": 9.651052631578948e-05,
|
11396 |
+
"loss": 0.0055,
|
11397 |
+
"step": 1663
|
11398 |
+
},
|
11399 |
+
{
|
11400 |
+
"epoch": 166.4,
|
11401 |
+
"learning_rate": 9.650526315789474e-05,
|
11402 |
+
"loss": 0.0061,
|
11403 |
+
"step": 1664
|
11404 |
+
},
|
11405 |
+
{
|
11406 |
+
"epoch": 166.5,
|
11407 |
+
"learning_rate": 9.65e-05,
|
11408 |
+
"loss": 0.0047,
|
11409 |
+
"step": 1665
|
11410 |
+
},
|
11411 |
+
{
|
11412 |
+
"epoch": 166.6,
|
11413 |
+
"learning_rate": 9.649473684210528e-05,
|
11414 |
+
"loss": 0.0046,
|
11415 |
+
"step": 1666
|
11416 |
+
},
|
11417 |
+
{
|
11418 |
+
"epoch": 166.7,
|
11419 |
+
"learning_rate": 9.648947368421053e-05,
|
11420 |
+
"loss": 0.0067,
|
11421 |
+
"step": 1667
|
11422 |
+
},
|
11423 |
+
{
|
11424 |
+
"epoch": 166.8,
|
11425 |
+
"learning_rate": 9.64842105263158e-05,
|
11426 |
+
"loss": 0.0063,
|
11427 |
+
"step": 1668
|
11428 |
+
},
|
11429 |
+
{
|
11430 |
+
"epoch": 166.9,
|
11431 |
+
"learning_rate": 9.647894736842106e-05,
|
11432 |
+
"loss": 0.0059,
|
11433 |
+
"step": 1669
|
11434 |
+
},
|
11435 |
+
{
|
11436 |
+
"epoch": 167.0,
|
11437 |
+
"learning_rate": 9.647368421052631e-05,
|
11438 |
+
"loss": 0.006,
|
11439 |
+
"step": 1670
|
11440 |
+
},
|
11441 |
+
{
|
11442 |
+
"epoch": 167.1,
|
11443 |
+
"learning_rate": 9.646842105263159e-05,
|
11444 |
+
"loss": 0.0061,
|
11445 |
+
"step": 1671
|
11446 |
+
},
|
11447 |
+
{
|
11448 |
+
"epoch": 167.2,
|
11449 |
+
"learning_rate": 9.646315789473685e-05,
|
11450 |
+
"loss": 0.0056,
|
11451 |
+
"step": 1672
|
11452 |
+
},
|
11453 |
+
{
|
11454 |
+
"epoch": 167.3,
|
11455 |
+
"learning_rate": 9.645789473684211e-05,
|
11456 |
+
"loss": 0.0061,
|
11457 |
+
"step": 1673
|
11458 |
+
},
|
11459 |
+
{
|
11460 |
+
"epoch": 167.4,
|
11461 |
+
"learning_rate": 9.645263157894738e-05,
|
11462 |
+
"loss": 0.0051,
|
11463 |
+
"step": 1674
|
11464 |
+
},
|
11465 |
+
{
|
11466 |
+
"epoch": 167.5,
|
11467 |
+
"learning_rate": 9.644736842105262e-05,
|
11468 |
+
"loss": 0.0052,
|
11469 |
+
"step": 1675
|
11470 |
+
},
|
11471 |
+
{
|
11472 |
+
"epoch": 167.6,
|
11473 |
+
"learning_rate": 9.64421052631579e-05,
|
11474 |
+
"loss": 0.0076,
|
11475 |
+
"step": 1676
|
11476 |
+
},
|
11477 |
+
{
|
11478 |
+
"epoch": 167.7,
|
11479 |
+
"learning_rate": 9.643684210526316e-05,
|
11480 |
+
"loss": 0.0057,
|
11481 |
+
"step": 1677
|
11482 |
+
},
|
11483 |
+
{
|
11484 |
+
"epoch": 167.8,
|
11485 |
+
"learning_rate": 9.643157894736842e-05,
|
11486 |
+
"loss": 0.0057,
|
11487 |
+
"step": 1678
|
11488 |
+
},
|
11489 |
+
{
|
11490 |
+
"epoch": 167.9,
|
11491 |
+
"learning_rate": 9.642631578947369e-05,
|
11492 |
+
"loss": 0.0052,
|
11493 |
+
"step": 1679
|
11494 |
+
},
|
11495 |
+
{
|
11496 |
+
"epoch": 168.0,
|
11497 |
+
"learning_rate": 9.642105263157896e-05,
|
11498 |
+
"loss": 0.0049,
|
11499 |
+
"step": 1680
|
11500 |
+
},
|
11501 |
+
{
|
11502 |
+
"epoch": 168.0,
|
11503 |
+
"eval_accuracy_safe": 0.8435761620040015,
|
11504 |
+
"eval_accuracy_unlabeled": NaN,
|
11505 |
+
"eval_accuracy_unsafe": 0.9961610096095911,
|
11506 |
+
"eval_iou_safe": 0.7490949734439962,
|
11507 |
+
"eval_iou_unlabeled": NaN,
|
11508 |
+
"eval_iou_unsafe": 0.9914406161065407,
|
11509 |
+
"eval_loss": 0.03941677510738373,
|
11510 |
+
"eval_mean_accuracy": 0.9198685858067963,
|
11511 |
+
"eval_mean_iou": 0.8702677947752684,
|
11512 |
+
"eval_overall_accuracy": 0.9916538978690532,
|
11513 |
+
"eval_runtime": 9.8722,
|
11514 |
+
"eval_samples_per_second": 6.787,
|
11515 |
+
"eval_steps_per_second": 0.506,
|
11516 |
+
"step": 1680
|
11517 |
+
},
|
11518 |
+
{
|
11519 |
+
"epoch": 168.1,
|
11520 |
+
"learning_rate": 9.641578947368421e-05,
|
11521 |
+
"loss": 0.0056,
|
11522 |
+
"step": 1681
|
11523 |
+
},
|
11524 |
+
{
|
11525 |
+
"epoch": 168.2,
|
11526 |
+
"learning_rate": 9.641052631578947e-05,
|
11527 |
+
"loss": 0.0063,
|
11528 |
+
"step": 1682
|
11529 |
+
},
|
11530 |
+
{
|
11531 |
+
"epoch": 168.3,
|
11532 |
+
"learning_rate": 9.640526315789474e-05,
|
11533 |
+
"loss": 0.0065,
|
11534 |
+
"step": 1683
|
11535 |
+
},
|
11536 |
+
{
|
11537 |
+
"epoch": 168.4,
|
11538 |
+
"learning_rate": 9.64e-05,
|
11539 |
+
"loss": 0.0056,
|
11540 |
+
"step": 1684
|
11541 |
+
},
|
11542 |
+
{
|
11543 |
+
"epoch": 168.5,
|
11544 |
+
"learning_rate": 9.639473684210527e-05,
|
11545 |
+
"loss": 0.0056,
|
11546 |
+
"step": 1685
|
11547 |
+
},
|
11548 |
+
{
|
11549 |
+
"epoch": 168.6,
|
11550 |
+
"learning_rate": 9.638947368421052e-05,
|
11551 |
+
"loss": 0.0046,
|
11552 |
+
"step": 1686
|
11553 |
+
},
|
11554 |
+
{
|
11555 |
+
"epoch": 168.7,
|
11556 |
+
"learning_rate": 9.63842105263158e-05,
|
11557 |
+
"loss": 0.0058,
|
11558 |
+
"step": 1687
|
11559 |
+
},
|
11560 |
+
{
|
11561 |
+
"epoch": 168.8,
|
11562 |
+
"learning_rate": 9.637894736842106e-05,
|
11563 |
+
"loss": 0.0045,
|
11564 |
+
"step": 1688
|
11565 |
+
},
|
11566 |
+
{
|
11567 |
+
"epoch": 168.9,
|
11568 |
+
"learning_rate": 9.637368421052631e-05,
|
11569 |
+
"loss": 0.0064,
|
11570 |
+
"step": 1689
|
11571 |
+
},
|
11572 |
+
{
|
11573 |
+
"epoch": 169.0,
|
11574 |
+
"learning_rate": 9.636842105263159e-05,
|
11575 |
+
"loss": 0.0056,
|
11576 |
+
"step": 1690
|
11577 |
+
},
|
11578 |
+
{
|
11579 |
+
"epoch": 169.1,
|
11580 |
+
"learning_rate": 9.636315789473685e-05,
|
11581 |
+
"loss": 0.0045,
|
11582 |
+
"step": 1691
|
11583 |
+
},
|
11584 |
+
{
|
11585 |
+
"epoch": 169.2,
|
11586 |
+
"learning_rate": 9.635789473684211e-05,
|
11587 |
+
"loss": 0.0072,
|
11588 |
+
"step": 1692
|
11589 |
+
},
|
11590 |
+
{
|
11591 |
+
"epoch": 169.3,
|
11592 |
+
"learning_rate": 9.635263157894737e-05,
|
11593 |
+
"loss": 0.0061,
|
11594 |
+
"step": 1693
|
11595 |
+
},
|
11596 |
+
{
|
11597 |
+
"epoch": 169.4,
|
11598 |
+
"learning_rate": 9.634736842105264e-05,
|
11599 |
+
"loss": 0.0054,
|
11600 |
+
"step": 1694
|
11601 |
+
},
|
11602 |
+
{
|
11603 |
+
"epoch": 169.5,
|
11604 |
+
"learning_rate": 9.63421052631579e-05,
|
11605 |
+
"loss": 0.0064,
|
11606 |
+
"step": 1695
|
11607 |
+
},
|
11608 |
+
{
|
11609 |
+
"epoch": 169.6,
|
11610 |
+
"learning_rate": 9.633684210526316e-05,
|
11611 |
+
"loss": 0.0053,
|
11612 |
+
"step": 1696
|
11613 |
+
},
|
11614 |
+
{
|
11615 |
+
"epoch": 169.7,
|
11616 |
+
"learning_rate": 9.633157894736842e-05,
|
11617 |
+
"loss": 0.0059,
|
11618 |
+
"step": 1697
|
11619 |
+
},
|
11620 |
+
{
|
11621 |
+
"epoch": 169.8,
|
11622 |
+
"learning_rate": 9.632631578947368e-05,
|
11623 |
+
"loss": 0.0052,
|
11624 |
+
"step": 1698
|
11625 |
+
},
|
11626 |
+
{
|
11627 |
+
"epoch": 169.9,
|
11628 |
+
"learning_rate": 9.632105263157896e-05,
|
11629 |
+
"loss": 0.0054,
|
11630 |
+
"step": 1699
|
11631 |
+
},
|
11632 |
+
{
|
11633 |
+
"epoch": 170.0,
|
11634 |
+
"learning_rate": 9.631578947368421e-05,
|
11635 |
+
"loss": 0.0054,
|
11636 |
+
"step": 1700
|
11637 |
+
},
|
11638 |
+
{
|
11639 |
+
"epoch": 170.0,
|
11640 |
+
"eval_accuracy_safe": 0.842791662329752,
|
11641 |
+
"eval_accuracy_unlabeled": NaN,
|
11642 |
+
"eval_accuracy_unsafe": 0.9962059498806853,
|
11643 |
+
"eval_iou_safe": 0.7493808604026593,
|
11644 |
+
"eval_iou_unlabeled": NaN,
|
11645 |
+
"eval_iou_unsafe": 0.991461781298423,
|
11646 |
+
"eval_loss": 0.03999360650777817,
|
11647 |
+
"eval_mean_accuracy": 0.9194988061052187,
|
11648 |
+
"eval_mean_iou": 0.8704213208505411,
|
11649 |
+
"eval_overall_accuracy": 0.9916743378141033,
|
11650 |
+
"eval_runtime": 9.8432,
|
11651 |
+
"eval_samples_per_second": 6.807,
|
11652 |
+
"eval_steps_per_second": 0.508,
|
11653 |
+
"step": 1700
|
11654 |
+
},
|
11655 |
+
{
|
11656 |
+
"epoch": 170.1,
|
11657 |
+
"learning_rate": 9.631052631578947e-05,
|
11658 |
+
"loss": 0.006,
|
11659 |
+
"step": 1701
|
11660 |
+
},
|
11661 |
+
{
|
11662 |
+
"epoch": 170.2,
|
11663 |
+
"learning_rate": 9.630526315789475e-05,
|
11664 |
+
"loss": 0.0057,
|
11665 |
+
"step": 1702
|
11666 |
+
},
|
11667 |
+
{
|
11668 |
+
"epoch": 170.3,
|
11669 |
+
"learning_rate": 9.63e-05,
|
11670 |
+
"loss": 0.0061,
|
11671 |
+
"step": 1703
|
11672 |
+
},
|
11673 |
+
{
|
11674 |
+
"epoch": 170.4,
|
11675 |
+
"learning_rate": 9.629473684210527e-05,
|
11676 |
+
"loss": 0.0053,
|
11677 |
+
"step": 1704
|
11678 |
+
},
|
11679 |
+
{
|
11680 |
+
"epoch": 170.5,
|
11681 |
+
"learning_rate": 9.628947368421053e-05,
|
11682 |
+
"loss": 0.0058,
|
11683 |
+
"step": 1705
|
11684 |
+
},
|
11685 |
+
{
|
11686 |
+
"epoch": 170.6,
|
11687 |
+
"learning_rate": 9.62842105263158e-05,
|
11688 |
+
"loss": 0.0045,
|
11689 |
+
"step": 1706
|
11690 |
+
},
|
11691 |
+
{
|
11692 |
+
"epoch": 170.7,
|
11693 |
+
"learning_rate": 9.627894736842106e-05,
|
11694 |
+
"loss": 0.0065,
|
11695 |
+
"step": 1707
|
11696 |
+
},
|
11697 |
+
{
|
11698 |
+
"epoch": 170.8,
|
11699 |
+
"learning_rate": 9.627368421052632e-05,
|
11700 |
+
"loss": 0.0055,
|
11701 |
+
"step": 1708
|
11702 |
+
},
|
11703 |
+
{
|
11704 |
+
"epoch": 170.9,
|
11705 |
+
"learning_rate": 9.626842105263158e-05,
|
11706 |
+
"loss": 0.0052,
|
11707 |
+
"step": 1709
|
11708 |
+
},
|
11709 |
+
{
|
11710 |
+
"epoch": 171.0,
|
11711 |
+
"learning_rate": 9.626315789473685e-05,
|
11712 |
+
"loss": 0.0056,
|
11713 |
+
"step": 1710
|
11714 |
+
},
|
11715 |
+
{
|
11716 |
+
"epoch": 171.1,
|
11717 |
+
"learning_rate": 9.625789473684211e-05,
|
11718 |
+
"loss": 0.0052,
|
11719 |
+
"step": 1711
|
11720 |
+
},
|
11721 |
+
{
|
11722 |
+
"epoch": 171.2,
|
11723 |
+
"learning_rate": 9.625263157894737e-05,
|
11724 |
+
"loss": 0.005,
|
11725 |
+
"step": 1712
|
11726 |
+
},
|
11727 |
+
{
|
11728 |
+
"epoch": 171.3,
|
11729 |
+
"learning_rate": 9.624736842105265e-05,
|
11730 |
+
"loss": 0.0051,
|
11731 |
+
"step": 1713
|
11732 |
+
},
|
11733 |
+
{
|
11734 |
+
"epoch": 171.4,
|
11735 |
+
"learning_rate": 9.62421052631579e-05,
|
11736 |
+
"loss": 0.0055,
|
11737 |
+
"step": 1714
|
11738 |
+
},
|
11739 |
+
{
|
11740 |
+
"epoch": 171.5,
|
11741 |
+
"learning_rate": 9.623684210526316e-05,
|
11742 |
+
"loss": 0.0065,
|
11743 |
+
"step": 1715
|
11744 |
+
},
|
11745 |
+
{
|
11746 |
+
"epoch": 171.6,
|
11747 |
+
"learning_rate": 9.623157894736843e-05,
|
11748 |
+
"loss": 0.0066,
|
11749 |
+
"step": 1716
|
11750 |
+
},
|
11751 |
+
{
|
11752 |
+
"epoch": 171.7,
|
11753 |
+
"learning_rate": 9.622631578947368e-05,
|
11754 |
+
"loss": 0.0051,
|
11755 |
+
"step": 1717
|
11756 |
+
},
|
11757 |
+
{
|
11758 |
+
"epoch": 171.8,
|
11759 |
+
"learning_rate": 9.622105263157896e-05,
|
11760 |
+
"loss": 0.006,
|
11761 |
+
"step": 1718
|
11762 |
+
},
|
11763 |
+
{
|
11764 |
+
"epoch": 171.9,
|
11765 |
+
"learning_rate": 9.621578947368422e-05,
|
11766 |
+
"loss": 0.0061,
|
11767 |
+
"step": 1719
|
11768 |
+
},
|
11769 |
+
{
|
11770 |
+
"epoch": 172.0,
|
11771 |
+
"learning_rate": 9.621052631578947e-05,
|
11772 |
+
"loss": 0.0046,
|
11773 |
+
"step": 1720
|
11774 |
+
},
|
11775 |
+
{
|
11776 |
+
"epoch": 172.0,
|
11777 |
+
"eval_accuracy_safe": 0.8409701581720965,
|
11778 |
+
"eval_accuracy_unlabeled": NaN,
|
11779 |
+
"eval_accuracy_unsafe": 0.9964832771149708,
|
11780 |
+
"eval_iou_safe": 0.753868722613107,
|
11781 |
+
"eval_iou_unlabeled": NaN,
|
11782 |
+
"eval_iou_unsafe": 0.991683068794468,
|
11783 |
+
"eval_loss": 0.03979964926838875,
|
11784 |
+
"eval_mean_accuracy": 0.9187267176435336,
|
11785 |
+
"eval_mean_iou": 0.8727758957037874,
|
11786 |
+
"eval_overall_accuracy": 0.9918896689343808,
|
11787 |
+
"eval_runtime": 9.9749,
|
11788 |
+
"eval_samples_per_second": 6.717,
|
11789 |
+
"eval_steps_per_second": 0.501,
|
11790 |
+
"step": 1720
|
11791 |
+
},
|
11792 |
+
{
|
11793 |
+
"epoch": 172.1,
|
11794 |
+
"learning_rate": 9.620526315789474e-05,
|
11795 |
+
"loss": 0.0071,
|
11796 |
+
"step": 1721
|
11797 |
+
},
|
11798 |
+
{
|
11799 |
+
"epoch": 172.2,
|
11800 |
+
"learning_rate": 9.620000000000001e-05,
|
11801 |
+
"loss": 0.0051,
|
11802 |
+
"step": 1722
|
11803 |
+
},
|
11804 |
+
{
|
11805 |
+
"epoch": 172.3,
|
11806 |
+
"learning_rate": 9.619473684210527e-05,
|
11807 |
+
"loss": 0.0063,
|
11808 |
+
"step": 1723
|
11809 |
+
},
|
11810 |
+
{
|
11811 |
+
"epoch": 172.4,
|
11812 |
+
"learning_rate": 9.618947368421053e-05,
|
11813 |
+
"loss": 0.0044,
|
11814 |
+
"step": 1724
|
11815 |
+
},
|
11816 |
+
{
|
11817 |
+
"epoch": 172.5,
|
11818 |
+
"learning_rate": 9.61842105263158e-05,
|
11819 |
+
"loss": 0.0049,
|
11820 |
+
"step": 1725
|
11821 |
+
},
|
11822 |
+
{
|
11823 |
+
"epoch": 172.6,
|
11824 |
+
"learning_rate": 9.617894736842106e-05,
|
11825 |
+
"loss": 0.0057,
|
11826 |
+
"step": 1726
|
11827 |
+
},
|
11828 |
+
{
|
11829 |
+
"epoch": 172.7,
|
11830 |
+
"learning_rate": 9.617368421052632e-05,
|
11831 |
+
"loss": 0.0059,
|
11832 |
+
"step": 1727
|
11833 |
+
},
|
11834 |
+
{
|
11835 |
+
"epoch": 172.8,
|
11836 |
+
"learning_rate": 9.616842105263158e-05,
|
11837 |
+
"loss": 0.0072,
|
11838 |
+
"step": 1728
|
11839 |
+
},
|
11840 |
+
{
|
11841 |
+
"epoch": 172.9,
|
11842 |
+
"learning_rate": 9.616315789473684e-05,
|
11843 |
+
"loss": 0.0051,
|
11844 |
+
"step": 1729
|
11845 |
+
},
|
11846 |
+
{
|
11847 |
+
"epoch": 173.0,
|
11848 |
+
"learning_rate": 9.615789473684212e-05,
|
11849 |
+
"loss": 0.0051,
|
11850 |
+
"step": 1730
|
11851 |
+
},
|
11852 |
+
{
|
11853 |
+
"epoch": 173.1,
|
11854 |
+
"learning_rate": 9.615263157894737e-05,
|
11855 |
+
"loss": 0.0059,
|
11856 |
+
"step": 1731
|
11857 |
+
},
|
11858 |
+
{
|
11859 |
+
"epoch": 173.2,
|
11860 |
+
"learning_rate": 9.614736842105264e-05,
|
11861 |
+
"loss": 0.0064,
|
11862 |
+
"step": 1732
|
11863 |
+
},
|
11864 |
+
{
|
11865 |
+
"epoch": 173.3,
|
11866 |
+
"learning_rate": 9.61421052631579e-05,
|
11867 |
+
"loss": 0.0049,
|
11868 |
+
"step": 1733
|
11869 |
+
},
|
11870 |
+
{
|
11871 |
+
"epoch": 173.4,
|
11872 |
+
"learning_rate": 9.613684210526316e-05,
|
11873 |
+
"loss": 0.0055,
|
11874 |
+
"step": 1734
|
11875 |
+
},
|
11876 |
+
{
|
11877 |
+
"epoch": 173.5,
|
11878 |
+
"learning_rate": 9.613157894736843e-05,
|
11879 |
+
"loss": 0.0056,
|
11880 |
+
"step": 1735
|
11881 |
+
},
|
11882 |
+
{
|
11883 |
+
"epoch": 173.6,
|
11884 |
+
"learning_rate": 9.61263157894737e-05,
|
11885 |
+
"loss": 0.0047,
|
11886 |
+
"step": 1736
|
11887 |
+
},
|
11888 |
+
{
|
11889 |
+
"epoch": 173.7,
|
11890 |
+
"learning_rate": 9.612105263157896e-05,
|
11891 |
+
"loss": 0.0052,
|
11892 |
+
"step": 1737
|
11893 |
+
},
|
11894 |
+
{
|
11895 |
+
"epoch": 173.8,
|
11896 |
+
"learning_rate": 9.611578947368422e-05,
|
11897 |
+
"loss": 0.005,
|
11898 |
+
"step": 1738
|
11899 |
+
},
|
11900 |
+
{
|
11901 |
+
"epoch": 173.9,
|
11902 |
+
"learning_rate": 9.611052631578947e-05,
|
11903 |
+
"loss": 0.0058,
|
11904 |
+
"step": 1739
|
11905 |
+
},
|
11906 |
+
{
|
11907 |
+
"epoch": 174.0,
|
11908 |
+
"learning_rate": 9.610526315789474e-05,
|
11909 |
+
"loss": 0.0058,
|
11910 |
+
"step": 1740
|
11911 |
+
},
|
11912 |
+
{
|
11913 |
+
"epoch": 174.0,
|
11914 |
+
"eval_accuracy_safe": 0.8366525186872834,
|
11915 |
+
"eval_accuracy_unlabeled": NaN,
|
11916 |
+
"eval_accuracy_unsafe": 0.9965137262020437,
|
11917 |
+
"eval_iou_safe": 0.7506714514246184,
|
11918 |
+
"eval_iou_unlabeled": NaN,
|
11919 |
+
"eval_iou_unsafe": 0.9915836869660227,
|
11920 |
+
"eval_loss": 0.04020100086927414,
|
11921 |
+
"eval_mean_accuracy": 0.9165831224446636,
|
11922 |
+
"eval_mean_iou": 0.8711275691953206,
|
11923 |
+
"eval_overall_accuracy": 0.9917916824568563,
|
11924 |
+
"eval_runtime": 9.7236,
|
11925 |
+
"eval_samples_per_second": 6.89,
|
11926 |
+
"eval_steps_per_second": 0.514,
|
11927 |
+
"step": 1740
|
11928 |
+
},
|
11929 |
+
{
|
11930 |
+
"epoch": 174.1,
|
11931 |
+
"learning_rate": 9.61e-05,
|
11932 |
+
"loss": 0.0058,
|
11933 |
+
"step": 1741
|
11934 |
+
},
|
11935 |
+
{
|
11936 |
+
"epoch": 174.2,
|
11937 |
+
"learning_rate": 9.609473684210527e-05,
|
11938 |
+
"loss": 0.0055,
|
11939 |
+
"step": 1742
|
11940 |
+
},
|
11941 |
+
{
|
11942 |
+
"epoch": 174.3,
|
11943 |
+
"learning_rate": 9.608947368421053e-05,
|
11944 |
+
"loss": 0.0055,
|
11945 |
+
"step": 1743
|
11946 |
+
},
|
11947 |
+
{
|
11948 |
+
"epoch": 174.4,
|
11949 |
+
"learning_rate": 9.60842105263158e-05,
|
11950 |
+
"loss": 0.0063,
|
11951 |
+
"step": 1744
|
11952 |
+
},
|
11953 |
+
{
|
11954 |
+
"epoch": 174.5,
|
11955 |
+
"learning_rate": 9.607894736842105e-05,
|
11956 |
+
"loss": 0.005,
|
11957 |
+
"step": 1745
|
11958 |
+
},
|
11959 |
+
{
|
11960 |
+
"epoch": 174.6,
|
11961 |
+
"learning_rate": 9.607368421052632e-05,
|
11962 |
+
"loss": 0.0047,
|
11963 |
+
"step": 1746
|
11964 |
+
},
|
11965 |
+
{
|
11966 |
+
"epoch": 174.7,
|
11967 |
+
"learning_rate": 9.606842105263158e-05,
|
11968 |
+
"loss": 0.0055,
|
11969 |
+
"step": 1747
|
11970 |
+
},
|
11971 |
+
{
|
11972 |
+
"epoch": 174.8,
|
11973 |
+
"learning_rate": 9.606315789473684e-05,
|
11974 |
+
"loss": 0.006,
|
11975 |
+
"step": 1748
|
11976 |
+
},
|
11977 |
+
{
|
11978 |
+
"epoch": 174.9,
|
11979 |
+
"learning_rate": 9.605789473684212e-05,
|
11980 |
+
"loss": 0.0049,
|
11981 |
+
"step": 1749
|
11982 |
+
},
|
11983 |
+
{
|
11984 |
+
"epoch": 175.0,
|
11985 |
+
"learning_rate": 9.605263157894737e-05,
|
11986 |
+
"loss": 0.0082,
|
11987 |
+
"step": 1750
|
11988 |
+
},
|
11989 |
+
{
|
11990 |
+
"epoch": 175.1,
|
11991 |
+
"learning_rate": 9.604736842105264e-05,
|
11992 |
+
"loss": 0.0057,
|
11993 |
+
"step": 1751
|
11994 |
+
},
|
11995 |
+
{
|
11996 |
+
"epoch": 175.2,
|
11997 |
+
"learning_rate": 9.60421052631579e-05,
|
11998 |
+
"loss": 0.0046,
|
11999 |
+
"step": 1752
|
12000 |
+
},
|
12001 |
+
{
|
12002 |
+
"epoch": 175.3,
|
12003 |
+
"learning_rate": 9.603684210526315e-05,
|
12004 |
+
"loss": 0.0053,
|
12005 |
+
"step": 1753
|
12006 |
+
},
|
12007 |
+
{
|
12008 |
+
"epoch": 175.4,
|
12009 |
+
"learning_rate": 9.603157894736843e-05,
|
12010 |
+
"loss": 0.0057,
|
12011 |
+
"step": 1754
|
12012 |
+
},
|
12013 |
+
{
|
12014 |
+
"epoch": 175.5,
|
12015 |
+
"learning_rate": 9.602631578947369e-05,
|
12016 |
+
"loss": 0.0067,
|
12017 |
+
"step": 1755
|
12018 |
+
},
|
12019 |
+
{
|
12020 |
+
"epoch": 175.6,
|
12021 |
+
"learning_rate": 9.602105263157895e-05,
|
12022 |
+
"loss": 0.0047,
|
12023 |
+
"step": 1756
|
12024 |
+
},
|
12025 |
+
{
|
12026 |
+
"epoch": 175.7,
|
12027 |
+
"learning_rate": 9.601578947368422e-05,
|
12028 |
+
"loss": 0.005,
|
12029 |
+
"step": 1757
|
12030 |
+
},
|
12031 |
+
{
|
12032 |
+
"epoch": 175.8,
|
12033 |
+
"learning_rate": 9.601052631578948e-05,
|
12034 |
+
"loss": 0.0063,
|
12035 |
+
"step": 1758
|
12036 |
+
},
|
12037 |
+
{
|
12038 |
+
"epoch": 175.9,
|
12039 |
+
"learning_rate": 9.600526315789474e-05,
|
12040 |
+
"loss": 0.006,
|
12041 |
+
"step": 1759
|
12042 |
+
},
|
12043 |
+
{
|
12044 |
+
"epoch": 176.0,
|
12045 |
+
"learning_rate": 9.6e-05,
|
12046 |
+
"loss": 0.005,
|
12047 |
+
"step": 1760
|
12048 |
+
},
|
12049 |
+
{
|
12050 |
+
"epoch": 176.0,
|
12051 |
+
"eval_accuracy_safe": 0.8427820247416162,
|
12052 |
+
"eval_accuracy_unlabeled": NaN,
|
12053 |
+
"eval_accuracy_unsafe": 0.9963479282828369,
|
12054 |
+
"eval_iou_safe": 0.7524933267475661,
|
12055 |
+
"eval_iou_unlabeled": NaN,
|
12056 |
+
"eval_iou_unsafe": 0.9916027940705149,
|
12057 |
+
"eval_loss": 0.0399840846657753,
|
12058 |
+
"eval_mean_accuracy": 0.9195649765122266,
|
12059 |
+
"eval_mean_iou": 0.8720480604090405,
|
12060 |
+
"eval_overall_accuracy": 0.9918118377230061,
|
12061 |
+
"eval_runtime": 9.5896,
|
12062 |
+
"eval_samples_per_second": 6.987,
|
12063 |
+
"eval_steps_per_second": 0.521,
|
12064 |
+
"step": 1760
|
12065 |
+
},
|
12066 |
+
{
|
12067 |
+
"epoch": 176.1,
|
12068 |
+
"learning_rate": 9.599473684210526e-05,
|
12069 |
+
"loss": 0.0055,
|
12070 |
+
"step": 1761
|
12071 |
+
},
|
12072 |
+
{
|
12073 |
+
"epoch": 176.2,
|
12074 |
+
"learning_rate": 9.598947368421053e-05,
|
12075 |
+
"loss": 0.0055,
|
12076 |
+
"step": 1762
|
12077 |
+
},
|
12078 |
+
{
|
12079 |
+
"epoch": 176.3,
|
12080 |
+
"learning_rate": 9.59842105263158e-05,
|
12081 |
+
"loss": 0.0068,
|
12082 |
+
"step": 1763
|
12083 |
+
},
|
12084 |
+
{
|
12085 |
+
"epoch": 176.4,
|
12086 |
+
"learning_rate": 9.597894736842105e-05,
|
12087 |
+
"loss": 0.0062,
|
12088 |
+
"step": 1764
|
12089 |
+
},
|
12090 |
+
{
|
12091 |
+
"epoch": 176.5,
|
12092 |
+
"learning_rate": 9.597368421052631e-05,
|
12093 |
+
"loss": 0.0062,
|
12094 |
+
"step": 1765
|
12095 |
+
},
|
12096 |
+
{
|
12097 |
+
"epoch": 176.6,
|
12098 |
+
"learning_rate": 9.596842105263159e-05,
|
12099 |
+
"loss": 0.005,
|
12100 |
+
"step": 1766
|
12101 |
+
},
|
12102 |
+
{
|
12103 |
+
"epoch": 176.7,
|
12104 |
+
"learning_rate": 9.596315789473684e-05,
|
12105 |
+
"loss": 0.0058,
|
12106 |
+
"step": 1767
|
12107 |
+
},
|
12108 |
+
{
|
12109 |
+
"epoch": 176.8,
|
12110 |
+
"learning_rate": 9.595789473684211e-05,
|
12111 |
+
"loss": 0.0048,
|
12112 |
+
"step": 1768
|
12113 |
+
},
|
12114 |
+
{
|
12115 |
+
"epoch": 176.9,
|
12116 |
+
"learning_rate": 9.595263157894738e-05,
|
12117 |
+
"loss": 0.0053,
|
12118 |
+
"step": 1769
|
12119 |
+
},
|
12120 |
+
{
|
12121 |
+
"epoch": 177.0,
|
12122 |
+
"learning_rate": 9.594736842105264e-05,
|
12123 |
+
"loss": 0.005,
|
12124 |
+
"step": 1770
|
12125 |
+
},
|
12126 |
+
{
|
12127 |
+
"epoch": 177.1,
|
12128 |
+
"learning_rate": 9.59421052631579e-05,
|
12129 |
+
"loss": 0.0059,
|
12130 |
+
"step": 1771
|
12131 |
+
},
|
12132 |
+
{
|
12133 |
+
"epoch": 177.2,
|
12134 |
+
"learning_rate": 9.593684210526316e-05,
|
12135 |
+
"loss": 0.0049,
|
12136 |
+
"step": 1772
|
12137 |
+
},
|
12138 |
+
{
|
12139 |
+
"epoch": 177.3,
|
12140 |
+
"learning_rate": 9.593157894736843e-05,
|
12141 |
+
"loss": 0.0055,
|
12142 |
+
"step": 1773
|
12143 |
+
},
|
12144 |
+
{
|
12145 |
+
"epoch": 177.4,
|
12146 |
+
"learning_rate": 9.592631578947369e-05,
|
12147 |
+
"loss": 0.007,
|
12148 |
+
"step": 1774
|
12149 |
+
},
|
12150 |
+
{
|
12151 |
+
"epoch": 177.5,
|
12152 |
+
"learning_rate": 9.592105263157895e-05,
|
12153 |
+
"loss": 0.0062,
|
12154 |
+
"step": 1775
|
12155 |
+
},
|
12156 |
+
{
|
12157 |
+
"epoch": 177.6,
|
12158 |
+
"learning_rate": 9.591578947368421e-05,
|
12159 |
+
"loss": 0.0051,
|
12160 |
+
"step": 1776
|
12161 |
+
},
|
12162 |
+
{
|
12163 |
+
"epoch": 177.7,
|
12164 |
+
"learning_rate": 9.591052631578948e-05,
|
12165 |
+
"loss": 0.0053,
|
12166 |
+
"step": 1777
|
12167 |
+
},
|
12168 |
+
{
|
12169 |
+
"epoch": 177.8,
|
12170 |
+
"learning_rate": 9.590526315789474e-05,
|
12171 |
+
"loss": 0.0049,
|
12172 |
+
"step": 1778
|
12173 |
+
},
|
12174 |
+
{
|
12175 |
+
"epoch": 177.9,
|
12176 |
+
"learning_rate": 9.59e-05,
|
12177 |
+
"loss": 0.0051,
|
12178 |
+
"step": 1779
|
12179 |
+
},
|
12180 |
+
{
|
12181 |
+
"epoch": 178.0,
|
12182 |
+
"learning_rate": 9.589473684210528e-05,
|
12183 |
+
"loss": 0.0061,
|
12184 |
+
"step": 1780
|
12185 |
+
},
|
12186 |
+
{
|
12187 |
+
"epoch": 178.0,
|
12188 |
+
"eval_accuracy_safe": 0.8491833107813771,
|
12189 |
+
"eval_accuracy_unlabeled": NaN,
|
12190 |
+
"eval_accuracy_unsafe": 0.9960346957666851,
|
12191 |
+
"eval_iou_safe": 0.7513054451815345,
|
12192 |
+
"eval_iou_unlabeled": NaN,
|
12193 |
+
"eval_iou_unsafe": 0.9914833128833639,
|
12194 |
+
"eval_loss": 0.04173571988940239,
|
12195 |
+
"eval_mean_accuracy": 0.9226090032740311,
|
12196 |
+
"eval_mean_iou": 0.8713943790324492,
|
12197 |
+
"eval_overall_accuracy": 0.9916969413187966,
|
12198 |
+
"eval_runtime": 9.6891,
|
12199 |
+
"eval_samples_per_second": 6.915,
|
12200 |
+
"eval_steps_per_second": 0.516,
|
12201 |
+
"step": 1780
|
12202 |
+
},
|
12203 |
+
{
|
12204 |
+
"epoch": 178.1,
|
12205 |
+
"learning_rate": 9.588947368421052e-05,
|
12206 |
+
"loss": 0.0059,
|
12207 |
+
"step": 1781
|
12208 |
+
},
|
12209 |
+
{
|
12210 |
+
"epoch": 178.2,
|
12211 |
+
"learning_rate": 9.58842105263158e-05,
|
12212 |
+
"loss": 0.006,
|
12213 |
+
"step": 1782
|
12214 |
+
},
|
12215 |
+
{
|
12216 |
+
"epoch": 178.3,
|
12217 |
+
"learning_rate": 9.587894736842106e-05,
|
12218 |
+
"loss": 0.0043,
|
12219 |
+
"step": 1783
|
12220 |
+
},
|
12221 |
+
{
|
12222 |
+
"epoch": 178.4,
|
12223 |
+
"learning_rate": 9.587368421052631e-05,
|
12224 |
+
"loss": 0.0054,
|
12225 |
+
"step": 1784
|
12226 |
+
},
|
12227 |
+
{
|
12228 |
+
"epoch": 178.5,
|
12229 |
+
"learning_rate": 9.586842105263159e-05,
|
12230 |
+
"loss": 0.0069,
|
12231 |
+
"step": 1785
|
12232 |
+
},
|
12233 |
+
{
|
12234 |
+
"epoch": 178.6,
|
12235 |
+
"learning_rate": 9.586315789473685e-05,
|
12236 |
+
"loss": 0.0051,
|
12237 |
+
"step": 1786
|
12238 |
+
},
|
12239 |
+
{
|
12240 |
+
"epoch": 178.7,
|
12241 |
+
"learning_rate": 9.585789473684211e-05,
|
12242 |
+
"loss": 0.0048,
|
12243 |
+
"step": 1787
|
12244 |
+
},
|
12245 |
+
{
|
12246 |
+
"epoch": 178.8,
|
12247 |
+
"learning_rate": 9.585263157894737e-05,
|
12248 |
+
"loss": 0.0054,
|
12249 |
+
"step": 1788
|
12250 |
+
},
|
12251 |
+
{
|
12252 |
+
"epoch": 178.9,
|
12253 |
+
"learning_rate": 9.584736842105264e-05,
|
12254 |
+
"loss": 0.0058,
|
12255 |
+
"step": 1789
|
12256 |
+
},
|
12257 |
+
{
|
12258 |
+
"epoch": 179.0,
|
12259 |
+
"learning_rate": 9.58421052631579e-05,
|
12260 |
+
"loss": 0.0053,
|
12261 |
+
"step": 1790
|
12262 |
+
},
|
12263 |
+
{
|
12264 |
+
"epoch": 179.1,
|
12265 |
+
"learning_rate": 9.583684210526316e-05,
|
12266 |
+
"loss": 0.0069,
|
12267 |
+
"step": 1791
|
12268 |
+
},
|
12269 |
+
{
|
12270 |
+
"epoch": 179.2,
|
12271 |
+
"learning_rate": 9.583157894736842e-05,
|
12272 |
+
"loss": 0.0052,
|
12273 |
+
"step": 1792
|
12274 |
+
},
|
12275 |
+
{
|
12276 |
+
"epoch": 179.3,
|
12277 |
+
"learning_rate": 9.582631578947369e-05,
|
12278 |
+
"loss": 0.0042,
|
12279 |
+
"step": 1793
|
12280 |
+
},
|
12281 |
+
{
|
12282 |
+
"epoch": 179.4,
|
12283 |
+
"learning_rate": 9.582105263157896e-05,
|
12284 |
+
"loss": 0.0053,
|
12285 |
+
"step": 1794
|
12286 |
+
},
|
12287 |
+
{
|
12288 |
+
"epoch": 179.5,
|
12289 |
+
"learning_rate": 9.581578947368421e-05,
|
12290 |
+
"loss": 0.0057,
|
12291 |
+
"step": 1795
|
12292 |
+
},
|
12293 |
+
{
|
12294 |
+
"epoch": 179.6,
|
12295 |
+
"learning_rate": 9.581052631578947e-05,
|
12296 |
+
"loss": 0.0044,
|
12297 |
+
"step": 1796
|
12298 |
+
},
|
12299 |
+
{
|
12300 |
+
"epoch": 179.7,
|
12301 |
+
"learning_rate": 9.580526315789475e-05,
|
12302 |
+
"loss": 0.0061,
|
12303 |
+
"step": 1797
|
12304 |
+
},
|
12305 |
+
{
|
12306 |
+
"epoch": 179.8,
|
12307 |
+
"learning_rate": 9.58e-05,
|
12308 |
+
"loss": 0.0059,
|
12309 |
+
"step": 1798
|
12310 |
+
},
|
12311 |
+
{
|
12312 |
+
"epoch": 179.9,
|
12313 |
+
"learning_rate": 9.579473684210527e-05,
|
12314 |
+
"loss": 0.0048,
|
12315 |
+
"step": 1799
|
12316 |
+
},
|
12317 |
+
{
|
12318 |
+
"epoch": 180.0,
|
12319 |
+
"learning_rate": 9.578947368421052e-05,
|
12320 |
+
"loss": 0.0061,
|
12321 |
+
"step": 1800
|
12322 |
+
},
|
12323 |
+
{
|
12324 |
+
"epoch": 180.0,
|
12325 |
+
"eval_accuracy_safe": 0.8537785128044996,
|
12326 |
+
"eval_accuracy_unlabeled": NaN,
|
12327 |
+
"eval_accuracy_unsafe": 0.9959957983779965,
|
12328 |
+
"eval_iou_safe": 0.7545179055383415,
|
12329 |
+
"eval_iou_unlabeled": NaN,
|
12330 |
+
"eval_iou_unsafe": 0.9915826485192842,
|
12331 |
+
"eval_loss": 0.04071439057588577,
|
12332 |
+
"eval_mean_accuracy": 0.924887155591248,
|
12333 |
+
"eval_mean_iou": 0.8730502770288129,
|
12334 |
+
"eval_overall_accuracy": 0.9917949277963212,
|
12335 |
+
"eval_runtime": 9.6123,
|
12336 |
+
"eval_samples_per_second": 6.97,
|
12337 |
+
"eval_steps_per_second": 0.52,
|
12338 |
+
"step": 1800
|
12339 |
+
},
|
12340 |
+
{
|
12341 |
+
"epoch": 180.1,
|
12342 |
+
"learning_rate": 9.57842105263158e-05,
|
12343 |
+
"loss": 0.0048,
|
12344 |
+
"step": 1801
|
12345 |
+
},
|
12346 |
+
{
|
12347 |
+
"epoch": 180.2,
|
12348 |
+
"learning_rate": 9.577894736842106e-05,
|
12349 |
+
"loss": 0.0048,
|
12350 |
+
"step": 1802
|
12351 |
+
},
|
12352 |
+
{
|
12353 |
+
"epoch": 180.3,
|
12354 |
+
"learning_rate": 9.577368421052631e-05,
|
12355 |
+
"loss": 0.0049,
|
12356 |
+
"step": 1803
|
12357 |
+
},
|
12358 |
+
{
|
12359 |
+
"epoch": 180.4,
|
12360 |
+
"learning_rate": 9.576842105263159e-05,
|
12361 |
+
"loss": 0.0055,
|
12362 |
+
"step": 1804
|
12363 |
+
},
|
12364 |
+
{
|
12365 |
+
"epoch": 180.5,
|
12366 |
+
"learning_rate": 9.576315789473685e-05,
|
12367 |
+
"loss": 0.0056,
|
12368 |
+
"step": 1805
|
12369 |
+
},
|
12370 |
+
{
|
12371 |
+
"epoch": 180.6,
|
12372 |
+
"learning_rate": 9.575789473684211e-05,
|
12373 |
+
"loss": 0.0052,
|
12374 |
+
"step": 1806
|
12375 |
+
},
|
12376 |
+
{
|
12377 |
+
"epoch": 180.7,
|
12378 |
+
"learning_rate": 9.575263157894737e-05,
|
12379 |
+
"loss": 0.005,
|
12380 |
+
"step": 1807
|
12381 |
+
},
|
12382 |
+
{
|
12383 |
+
"epoch": 180.8,
|
12384 |
+
"learning_rate": 9.574736842105265e-05,
|
12385 |
+
"loss": 0.005,
|
12386 |
+
"step": 1808
|
12387 |
+
},
|
12388 |
+
{
|
12389 |
+
"epoch": 180.9,
|
12390 |
+
"learning_rate": 9.57421052631579e-05,
|
12391 |
+
"loss": 0.0072,
|
12392 |
+
"step": 1809
|
12393 |
+
},
|
12394 |
+
{
|
12395 |
+
"epoch": 181.0,
|
12396 |
+
"learning_rate": 9.573684210526316e-05,
|
12397 |
+
"loss": 0.0057,
|
12398 |
+
"step": 1810
|
12399 |
+
},
|
12400 |
+
{
|
12401 |
+
"epoch": 181.1,
|
12402 |
+
"learning_rate": 9.573157894736842e-05,
|
12403 |
+
"loss": 0.0046,
|
12404 |
+
"step": 1811
|
12405 |
+
},
|
12406 |
+
{
|
12407 |
+
"epoch": 181.2,
|
12408 |
+
"learning_rate": 9.572631578947368e-05,
|
12409 |
+
"loss": 0.0059,
|
12410 |
+
"step": 1812
|
12411 |
+
},
|
12412 |
+
{
|
12413 |
+
"epoch": 181.3,
|
12414 |
+
"learning_rate": 9.572105263157896e-05,
|
12415 |
+
"loss": 0.006,
|
12416 |
+
"step": 1813
|
12417 |
+
},
|
12418 |
+
{
|
12419 |
+
"epoch": 181.4,
|
12420 |
+
"learning_rate": 9.571578947368421e-05,
|
12421 |
+
"loss": 0.0047,
|
12422 |
+
"step": 1814
|
12423 |
+
},
|
12424 |
+
{
|
12425 |
+
"epoch": 181.5,
|
12426 |
+
"learning_rate": 9.571052631578948e-05,
|
12427 |
+
"loss": 0.0053,
|
12428 |
+
"step": 1815
|
12429 |
+
},
|
12430 |
+
{
|
12431 |
+
"epoch": 181.6,
|
12432 |
+
"learning_rate": 9.570526315789475e-05,
|
12433 |
+
"loss": 0.0065,
|
12434 |
+
"step": 1816
|
12435 |
+
},
|
12436 |
+
{
|
12437 |
+
"epoch": 181.7,
|
12438 |
+
"learning_rate": 9.57e-05,
|
12439 |
+
"loss": 0.0048,
|
12440 |
+
"step": 1817
|
12441 |
+
},
|
12442 |
+
{
|
12443 |
+
"epoch": 181.8,
|
12444 |
+
"learning_rate": 9.569473684210527e-05,
|
12445 |
+
"loss": 0.0045,
|
12446 |
+
"step": 1818
|
12447 |
+
},
|
12448 |
+
{
|
12449 |
+
"epoch": 181.9,
|
12450 |
+
"learning_rate": 9.568947368421053e-05,
|
12451 |
+
"loss": 0.0049,
|
12452 |
+
"step": 1819
|
12453 |
+
},
|
12454 |
+
{
|
12455 |
+
"epoch": 182.0,
|
12456 |
+
"learning_rate": 9.56842105263158e-05,
|
12457 |
+
"loss": 0.0065,
|
12458 |
+
"step": 1820
|
12459 |
+
},
|
12460 |
+
{
|
12461 |
+
"epoch": 182.0,
|
12462 |
+
"eval_accuracy_safe": 0.8510607129502199,
|
12463 |
+
"eval_accuracy_unlabeled": NaN,
|
12464 |
+
"eval_accuracy_unsafe": 0.9959411190925398,
|
12465 |
+
"eval_iou_safe": 0.7509239181271631,
|
12466 |
+
"eval_iou_unlabeled": NaN,
|
12467 |
+
"eval_iou_unsafe": 0.9914465595257035,
|
12468 |
+
"eval_loss": 0.04199281707406044,
|
12469 |
+
"eval_mean_accuracy": 0.9235009160213798,
|
12470 |
+
"eval_mean_iou": 0.8711852388264333,
|
12471 |
+
"eval_overall_accuracy": 0.9916615841993645,
|
12472 |
+
"eval_runtime": 9.9031,
|
12473 |
+
"eval_samples_per_second": 6.766,
|
12474 |
+
"eval_steps_per_second": 0.505,
|
12475 |
+
"step": 1820
|
12476 |
+
},
|
12477 |
+
{
|
12478 |
+
"epoch": 182.1,
|
12479 |
+
"learning_rate": 9.567894736842106e-05,
|
12480 |
+
"loss": 0.0054,
|
12481 |
+
"step": 1821
|
12482 |
+
},
|
12483 |
+
{
|
12484 |
+
"epoch": 182.2,
|
12485 |
+
"learning_rate": 9.567368421052632e-05,
|
12486 |
+
"loss": 0.0056,
|
12487 |
+
"step": 1822
|
12488 |
+
},
|
12489 |
+
{
|
12490 |
+
"epoch": 182.3,
|
12491 |
+
"learning_rate": 9.566842105263158e-05,
|
12492 |
+
"loss": 0.0051,
|
12493 |
+
"step": 1823
|
12494 |
+
},
|
12495 |
+
{
|
12496 |
+
"epoch": 182.4,
|
12497 |
+
"learning_rate": 9.566315789473684e-05,
|
12498 |
+
"loss": 0.0057,
|
12499 |
+
"step": 1824
|
12500 |
+
},
|
12501 |
+
{
|
12502 |
+
"epoch": 182.5,
|
12503 |
+
"learning_rate": 9.565789473684211e-05,
|
12504 |
+
"loss": 0.0055,
|
12505 |
+
"step": 1825
|
12506 |
+
},
|
12507 |
+
{
|
12508 |
+
"epoch": 182.6,
|
12509 |
+
"learning_rate": 9.565263157894737e-05,
|
12510 |
+
"loss": 0.0054,
|
12511 |
+
"step": 1826
|
12512 |
+
},
|
12513 |
+
{
|
12514 |
+
"epoch": 182.7,
|
12515 |
+
"learning_rate": 9.564736842105265e-05,
|
12516 |
+
"loss": 0.0054,
|
12517 |
+
"step": 1827
|
12518 |
+
},
|
12519 |
+
{
|
12520 |
+
"epoch": 182.8,
|
12521 |
+
"learning_rate": 9.56421052631579e-05,
|
12522 |
+
"loss": 0.005,
|
12523 |
+
"step": 1828
|
12524 |
+
},
|
12525 |
+
{
|
12526 |
+
"epoch": 182.9,
|
12527 |
+
"learning_rate": 9.563684210526316e-05,
|
12528 |
+
"loss": 0.0053,
|
12529 |
+
"step": 1829
|
12530 |
+
},
|
12531 |
+
{
|
12532 |
+
"epoch": 183.0,
|
12533 |
+
"learning_rate": 9.563157894736843e-05,
|
12534 |
+
"loss": 0.005,
|
12535 |
+
"step": 1830
|
12536 |
+
},
|
12537 |
+
{
|
12538 |
+
"epoch": 183.1,
|
12539 |
+
"learning_rate": 9.562631578947368e-05,
|
12540 |
+
"loss": 0.0052,
|
12541 |
+
"step": 1831
|
12542 |
+
},
|
12543 |
+
{
|
12544 |
+
"epoch": 183.2,
|
12545 |
+
"learning_rate": 9.562105263157896e-05,
|
12546 |
+
"loss": 0.0045,
|
12547 |
+
"step": 1832
|
12548 |
+
},
|
12549 |
+
{
|
12550 |
+
"epoch": 183.3,
|
12551 |
+
"learning_rate": 9.561578947368422e-05,
|
12552 |
+
"loss": 0.0056,
|
12553 |
+
"step": 1833
|
12554 |
+
},
|
12555 |
+
{
|
12556 |
+
"epoch": 183.4,
|
12557 |
+
"learning_rate": 9.561052631578948e-05,
|
12558 |
+
"loss": 0.0047,
|
12559 |
+
"step": 1834
|
12560 |
+
},
|
12561 |
+
{
|
12562 |
+
"epoch": 183.5,
|
12563 |
+
"learning_rate": 9.560526315789474e-05,
|
12564 |
+
"loss": 0.0047,
|
12565 |
+
"step": 1835
|
12566 |
+
},
|
12567 |
+
{
|
12568 |
+
"epoch": 183.6,
|
12569 |
+
"learning_rate": 9.56e-05,
|
12570 |
+
"loss": 0.0053,
|
12571 |
+
"step": 1836
|
12572 |
+
},
|
12573 |
+
{
|
12574 |
+
"epoch": 183.7,
|
12575 |
+
"learning_rate": 9.559473684210527e-05,
|
12576 |
+
"loss": 0.0063,
|
12577 |
+
"step": 1837
|
12578 |
+
},
|
12579 |
+
{
|
12580 |
+
"epoch": 183.8,
|
12581 |
+
"learning_rate": 9.558947368421053e-05,
|
12582 |
+
"loss": 0.005,
|
12583 |
+
"step": 1838
|
12584 |
+
},
|
12585 |
+
{
|
12586 |
+
"epoch": 183.9,
|
12587 |
+
"learning_rate": 9.558421052631579e-05,
|
12588 |
+
"loss": 0.006,
|
12589 |
+
"step": 1839
|
12590 |
+
},
|
12591 |
+
{
|
12592 |
+
"epoch": 184.0,
|
12593 |
+
"learning_rate": 9.557894736842106e-05,
|
12594 |
+
"loss": 0.0045,
|
12595 |
+
"step": 1840
|
12596 |
+
},
|
12597 |
+
{
|
12598 |
+
"epoch": 184.0,
|
12599 |
+
"eval_accuracy_safe": 0.854119683424505,
|
12600 |
+
"eval_accuracy_unlabeled": NaN,
|
12601 |
+
"eval_accuracy_unsafe": 0.9958763487801533,
|
12602 |
+
"eval_iou_safe": 0.7522106167319369,
|
12603 |
+
"eval_iou_unlabeled": NaN,
|
12604 |
+
"eval_iou_unsafe": 0.9914739784033543,
|
12605 |
+
"eval_loss": 0.04210684448480606,
|
12606 |
+
"eval_mean_accuracy": 0.9249980161023291,
|
12607 |
+
"eval_mean_iou": 0.8718422975676456,
|
12608 |
+
"eval_overall_accuracy": 0.9916890841811451,
|
12609 |
+
"eval_runtime": 9.7369,
|
12610 |
+
"eval_samples_per_second": 6.881,
|
12611 |
+
"eval_steps_per_second": 0.514,
|
12612 |
+
"step": 1840
|
12613 |
}
|
12614 |
],
|
12615 |
"max_steps": 20000,
|
12616 |
"num_train_epochs": 2000,
|
12617 |
+
"total_flos": 1.1310948050861556e+19,
|
12618 |
"trial_name": null,
|
12619 |
"trial_params": null
|
12620 |
}
|
{checkpoint-1520 β checkpoint-1840}/training_args.bin
RENAMED
File without changes
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 338796489
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42e1e3be6e685353b236ae09442bc140dfb5e35929e2ca289bd8856ee6accc20
|
3 |
size 338796489
|
runs/Feb08_09-49-29_robovision.csres.utexas.edu/events.out.tfevents.1707407376.robovision.csres.utexas.edu.889113.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b7f6aaa04920436597ad23a38eac50e10f426db46c74b401ef237a6db922986
|
3 |
+
size 365077
|