Training in progress, epoch 112, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 166496880
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3877cbb6b950cf4c5843dd0231527a1cc2b3b9bfc2562ccb8c27b7f9627eee3e
|
3 |
size 166496880
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 330495866
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92131f5675c8cd4e58cc1ba0f6c595c49b93168989163839bb1a382acb0fcadc
|
3 |
size 330495866
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c1fc25d09e52d766f0206907f84e9120f0722c8875148317201f5f9e6b824ba3
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26f9d12f584bdbd40d1e891beaa1bdd169a81c278721be8ba6ce2a5a585b5ca0
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.2962397336959839,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -15843,6 +15843,144 @@
|
|
15843 |
"eval_samples_per_second": 8.455,
|
15844 |
"eval_steps_per_second": 1.099,
|
15845 |
"step": 55500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15846 |
}
|
15847 |
],
|
15848 |
"logging_steps": 30,
|
@@ -15862,7 +16000,7 @@
|
|
15862 |
"attributes": {}
|
15863 |
}
|
15864 |
},
|
15865 |
-
"total_flos": 1.
|
15866 |
"train_batch_size": 2,
|
15867 |
"trial_name": null,
|
15868 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.2962397336959839,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
|
4 |
+
"epoch": 112.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 56000,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
15843 |
"eval_samples_per_second": 8.455,
|
15844 |
"eval_steps_per_second": 1.099,
|
15845 |
"step": 55500
|
15846 |
+
},
|
15847 |
+
{
|
15848 |
+
"epoch": 111.06,
|
15849 |
+
"grad_norm": 54.1436882019043,
|
15850 |
+
"learning_rate": 1.572686929076116e-06,
|
15851 |
+
"loss": 0.2357,
|
15852 |
+
"step": 55530
|
15853 |
+
},
|
15854 |
+
{
|
15855 |
+
"epoch": 111.12,
|
15856 |
+
"grad_norm": 107.09626007080078,
|
15857 |
+
"learning_rate": 1.5681147999908308e-06,
|
15858 |
+
"loss": 0.2024,
|
15859 |
+
"step": 55560
|
15860 |
+
},
|
15861 |
+
{
|
15862 |
+
"epoch": 111.18,
|
15863 |
+
"grad_norm": 40.910980224609375,
|
15864 |
+
"learning_rate": 1.5635480903207139e-06,
|
15865 |
+
"loss": 0.2048,
|
15866 |
+
"step": 55590
|
15867 |
+
},
|
15868 |
+
{
|
15869 |
+
"epoch": 111.24,
|
15870 |
+
"grad_norm": 76.96800994873047,
|
15871 |
+
"learning_rate": 1.5589868072772279e-06,
|
15872 |
+
"loss": 0.2338,
|
15873 |
+
"step": 55620
|
15874 |
+
},
|
15875 |
+
{
|
15876 |
+
"epoch": 111.3,
|
15877 |
+
"grad_norm": 77.27308654785156,
|
15878 |
+
"learning_rate": 1.554430958063259e-06,
|
15879 |
+
"loss": 0.2305,
|
15880 |
+
"step": 55650
|
15881 |
+
},
|
15882 |
+
{
|
15883 |
+
"epoch": 111.36,
|
15884 |
+
"grad_norm": 69.03377532958984,
|
15885 |
+
"learning_rate": 1.5498805498731146e-06,
|
15886 |
+
"loss": 0.2088,
|
15887 |
+
"step": 55680
|
15888 |
+
},
|
15889 |
+
{
|
15890 |
+
"epoch": 111.42,
|
15891 |
+
"grad_norm": 39.929359436035156,
|
15892 |
+
"learning_rate": 1.5453355898925094e-06,
|
15893 |
+
"loss": 0.2164,
|
15894 |
+
"step": 55710
|
15895 |
+
},
|
15896 |
+
{
|
15897 |
+
"epoch": 111.48,
|
15898 |
+
"grad_norm": 98.23054504394531,
|
15899 |
+
"learning_rate": 1.5407960852985582e-06,
|
15900 |
+
"loss": 0.2294,
|
15901 |
+
"step": 55740
|
15902 |
+
},
|
15903 |
+
{
|
15904 |
+
"epoch": 111.54,
|
15905 |
+
"grad_norm": 41.308589935302734,
|
15906 |
+
"learning_rate": 1.5362620432597559e-06,
|
15907 |
+
"loss": 0.2172,
|
15908 |
+
"step": 55770
|
15909 |
+
},
|
15910 |
+
{
|
15911 |
+
"epoch": 111.6,
|
15912 |
+
"grad_norm": 42.228031158447266,
|
15913 |
+
"learning_rate": 1.531733470935976e-06,
|
15914 |
+
"loss": 0.2229,
|
15915 |
+
"step": 55800
|
15916 |
+
},
|
15917 |
+
{
|
15918 |
+
"epoch": 111.66,
|
15919 |
+
"grad_norm": 44.37253189086914,
|
15920 |
+
"learning_rate": 1.5272103754784517e-06,
|
15921 |
+
"loss": 0.2366,
|
15922 |
+
"step": 55830
|
15923 |
+
},
|
15924 |
+
{
|
15925 |
+
"epoch": 111.72,
|
15926 |
+
"grad_norm": 59.920867919921875,
|
15927 |
+
"learning_rate": 1.5226927640297663e-06,
|
15928 |
+
"loss": 0.2056,
|
15929 |
+
"step": 55860
|
15930 |
+
},
|
15931 |
+
{
|
15932 |
+
"epoch": 111.78,
|
15933 |
+
"grad_norm": 103.45928955078125,
|
15934 |
+
"learning_rate": 1.5181806437238472e-06,
|
15935 |
+
"loss": 0.2811,
|
15936 |
+
"step": 55890
|
15937 |
+
},
|
15938 |
+
{
|
15939 |
+
"epoch": 111.84,
|
15940 |
+
"grad_norm": 31.924144744873047,
|
15941 |
+
"learning_rate": 1.5136740216859464e-06,
|
15942 |
+
"loss": 0.2208,
|
15943 |
+
"step": 55920
|
15944 |
+
},
|
15945 |
+
{
|
15946 |
+
"epoch": 111.9,
|
15947 |
+
"grad_norm": 34.80758285522461,
|
15948 |
+
"learning_rate": 1.5091729050326376e-06,
|
15949 |
+
"loss": 0.2034,
|
15950 |
+
"step": 55950
|
15951 |
+
},
|
15952 |
+
{
|
15953 |
+
"epoch": 111.96,
|
15954 |
+
"grad_norm": 77.86346435546875,
|
15955 |
+
"learning_rate": 1.5046773008717968e-06,
|
15956 |
+
"loss": 0.2089,
|
15957 |
+
"step": 55980
|
15958 |
+
},
|
15959 |
+
{
|
15960 |
+
"epoch": 112.0,
|
15961 |
+
"eval_loss": 0.31037819385528564,
|
15962 |
+
"eval_map": 0.8021,
|
15963 |
+
"eval_map_50": 0.9514,
|
15964 |
+
"eval_map_75": 0.9063,
|
15965 |
+
"eval_map_chicken": 0.8105,
|
15966 |
+
"eval_map_duck": 0.7278,
|
15967 |
+
"eval_map_large": 0.8079,
|
15968 |
+
"eval_map_medium": 0.8046,
|
15969 |
+
"eval_map_plant": 0.868,
|
15970 |
+
"eval_map_small": 0.3474,
|
15971 |
+
"eval_mar_1": 0.3175,
|
15972 |
+
"eval_mar_10": 0.8374,
|
15973 |
+
"eval_mar_100": 0.8407,
|
15974 |
+
"eval_mar_100_chicken": 0.8472,
|
15975 |
+
"eval_mar_100_duck": 0.7763,
|
15976 |
+
"eval_mar_100_plant": 0.8985,
|
15977 |
+
"eval_mar_large": 0.8355,
|
15978 |
+
"eval_mar_medium": 0.8471,
|
15979 |
+
"eval_mar_small": 0.3886,
|
15980 |
+
"eval_runtime": 11.3085,
|
15981 |
+
"eval_samples_per_second": 8.843,
|
15982 |
+
"eval_steps_per_second": 1.15,
|
15983 |
+
"step": 56000
|
15984 |
}
|
15985 |
],
|
15986 |
"logging_steps": 30,
|
|
|
16000 |
"attributes": {}
|
16001 |
}
|
16002 |
},
|
16003 |
+
"total_flos": 1.92652943228928e+19,
|
16004 |
"train_batch_size": 2,
|
16005 |
"trial_name": null,
|
16006 |
"trial_params": null
|