diff --git a/boxoban_data.h5 b/boxoban_data.h5 index dcfd6b4526de731e9a97cab372e9e45c3cee97de..2c84ae2d02f8eafdbc47bf69f8530611a4a448c8 100644 --- a/boxoban_data.h5 +++ b/boxoban_data.h5 @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8308df02af3dc4685d9b43454e43b6a623e53a6985436301450e757ed9f0e721 -size 2948509688 +oid sha256:2c2f3a6add7f182b6b39a27c34c3015ef83a21cc64871d5bcee8bd6a521cd7fa +size 3288195744 diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/config.json b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/global_step.txt b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..f418ce83668378a8d81057f8219a5fea544bb5ab --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/global_step.txt @@ -0,0 +1 @@ +78400 \ No newline at end of file diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/optimizer.pt b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..513f84ca2e742b38a2c45c7a3edc256f18449146 --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d62ac04f24947a0676763bd1937ea926bc0c1de8c574bc026307912f405df56 +size 995653957 diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/pytorch_model.bin b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c41c588386e08dc4cb8b56006af87dd876804e72 --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-78400/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c47508bdfe7c03e38eeb4f2ba94089736b09fb7e0f624c0b8075eaff08b25b80 +size 510404157 diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f2dca950a3113b4e289110750de13193a40370f2 --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json @@ -0,0 +1 @@ +{"game": "l_maze", "data_source": "l_maze", "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 40, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0} \ No newline at end of file diff --git a/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0289.3060598.0 b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0289.3060598.0 new file mode 100644 index 0000000000000000000000000000000000000000..3332f98ab4329d14c15a9d60607c3c8a3f1b1baf --- /dev/null +++ b/logs/l_maze copy/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0289.3060598.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53461354b466fd25f6c9f5ecb4300dd2b15a4e8cd8ad9ee1f354778cc74b4a8b +size 3983113 diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/config.json b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/config.json new file mode 100644 index 0000000000000000000000000000000000000000..be15b2ed9b3c86b42e3173b47610c0ea40e1507d --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/config.json @@ -0,0 +1,40 @@ +{ + "_name_or_path": "microsoft/CodeGPT-small-java-adaptedGPT2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "gradient_checkpointing": false, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50263 +} diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/global_step.txt b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..93ca626eaf9807762cb23e4ba59712ffdd1006c8 --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/global_step.txt @@ -0,0 +1 @@ +76160 \ No newline at end of file diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/optimizer.pt b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2283972dc43c36a624994a778f8198d92bd8d57f --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b4a4aa9a820a7ddd913240cdfb0c75199d5d35628722e3b60586dd623eb6582 +size 995678533 diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/pytorch_model.bin b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..07e391db8663f5d9bb2dbdd955d15ee0ea8d83a5 --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/checkpoint-76160/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b08dbc9363ef0f918442a9a62a61dafc026809ab18e44d0d05c22be3002ccc0 +size 510416445 diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7b4cff77f288554e6d6083f5809553369bd55deb --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/config.json @@ -0,0 +1 @@ +{"game": "l_maze", "data_source": "l_maze", "chunk_size": 128, "model": "java-gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 40, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0} \ No newline at end of file diff --git a/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0410.1063561.0 b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0410.1063561.0 new file mode 100644 index 0000000000000000000000000000000000000000..3155bc4f8437fc5207dbbd08f020ad756e6b3499 --- /dev/null +++ b/logs/l_maze copy/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1 copy/events.out.tfevents.1674499600.learnfair0410.1063561.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1a8b52a0a069dc516bb7aa26c87e7bc9b379f3ce0d3561048b44336d870b6d3 +size 3869180 diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/config.json b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/global_step.txt b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..f418ce83668378a8d81057f8219a5fea544bb5ab --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/global_step.txt @@ -0,0 +1 @@ +78400 \ No newline at end of file diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/optimizer.pt b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..513f84ca2e742b38a2c45c7a3edc256f18449146 --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d62ac04f24947a0676763bd1937ea926bc0c1de8c574bc026307912f405df56 +size 995653957 diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/pytorch_model.bin b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..c41c588386e08dc4cb8b56006af87dd876804e72 --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-78400/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c47508bdfe7c03e38eeb4f2ba94089736b09fb7e0f624c0b8075eaff08b25b80 +size 510404157 diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5bd5c6a4646af785465eac1085a7d00b3644027e --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json @@ -0,0 +1 @@ +{"game": "l_maze", "data_source": "l_maze", "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0} \ No newline at end of file diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0289.3060598.0 b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0289.3060598.0 new file mode 100644 index 0000000000000000000000000000000000000000..3332f98ab4329d14c15a9d60607c3c8a3f1b1baf --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0289.3060598.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53461354b466fd25f6c9f5ecb4300dd2b15a4e8cd8ad9ee1f354778cc74b4a8b +size 3983113 diff --git a/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674595502.devfair0748.3640277.0 b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674595502.devfair0748.3640277.0 new file mode 100644 index 0000000000000000000000000000000000000000..a412299ac388819e5422f7ba6eba7c4a3da49d99 --- /dev/null +++ b/logs/l_maze/gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674595502.devfair0748.3640277.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd8b5bc1662f04e442d54f48cfc18ff5259e8631820bc5beea7ddc0987dc10a8 +size 7165 diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/config.json b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/config.json new file mode 100644 index 0000000000000000000000000000000000000000..be15b2ed9b3c86b42e3173b47610c0ea40e1507d --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/config.json @@ -0,0 +1,40 @@ +{ + "_name_or_path": "microsoft/CodeGPT-small-java-adaptedGPT2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "gradient_checkpointing": false, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50263 +} diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/global_step.txt b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..93ca626eaf9807762cb23e4ba59712ffdd1006c8 --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/global_step.txt @@ -0,0 +1 @@ +76160 \ No newline at end of file diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/optimizer.pt b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2283972dc43c36a624994a778f8198d92bd8d57f --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b4a4aa9a820a7ddd913240cdfb0c75199d5d35628722e3b60586dd623eb6582 +size 995678533 diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/pytorch_model.bin b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..07e391db8663f5d9bb2dbdd955d15ee0ea8d83a5 --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/checkpoint-76160/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b08dbc9363ef0f918442a9a62a61dafc026809ab18e44d0d05c22be3002ccc0 +size 510416445 diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7b4cff77f288554e6d6083f5809553369bd55deb --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/config.json @@ -0,0 +1 @@ +{"game": "l_maze", "data_source": "l_maze", "chunk_size": 128, "model": "java-gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 40, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0} \ No newline at end of file diff --git a/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0410.1063561.0 b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0410.1063561.0 new file mode 100644 index 0000000000000000000000000000000000000000..3155bc4f8437fc5207dbbd08f020ad756e6b3499 --- /dev/null +++ b/logs/l_maze/java-gpt2/l_maze/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674499600.learnfair0410.1063561.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1a8b52a0a069dc516bb7aa26c87e7bc9b379f3ce0d3561048b44336d870b6d3 +size 3869180 diff --git a/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json b/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f2f330bc9027cc2b16b6d19df735a1c5384da6ea --- /dev/null +++ b/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "codeparrot", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": true, "seed": 42, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674871152.devfair0748.475905.0 b/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674871152.devfair0748.475905.0 new file mode 100644 index 0000000000000000000000000000000000000000..ed08b9ea4ea99d4cd97d242f4f7dbfa6273aac58 --- /dev/null +++ b/logs/sokoban/model:codeparrot/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674871152.devfair0748.475905.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33bc448370b7a4680ef47e9e6b79955ff07d10444e7f542ebd3d3b6691f6580d +size 88 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..57fa89ccd6bcaa6335cb8d6171238885fb96d771 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87d510c448635538a0843ab08dc6377517c76b397e8747e83d466ac232550301 +size 995653957 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..e4f8fccbad7e9be77ba5b05eacfbc0e9e72068a7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc124a986a9ba15f261978e1d7f7878fe726c273f12845858c5c694d8d1481ca +size 510404157 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json new file mode 100644 index 0000000000000000000000000000000000000000..5893c700287989f0da0640bb18550bd963629cac --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 0, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873092.learnfair0316.111781.0 b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873092.learnfair0316.111781.0 new file mode 100644 index 0000000000000000000000000000000000000000..93dc8a9376c9c3198f3532974e053b78937120a7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873092.learnfair0316.111781.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86367ac59d66b13e054e69aa375b3a8ada73ec6e725fe17c2e9ce369f07e44a4 +size 49424 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0c410de1f602dd652ffc7b7f77a9d911a76a1b2c --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:668e8ef6aef46cf04f2bead38ee89fdd9786f4f024727ce36faa6b6e44c8e3ed +size 995653957 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..02299ef638f77a8b128df42bbd3ccb22c3862ff1 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be0c2480d646c7fabc94d5e1c4c1caee523af08e29a2c0ab6d816e225805705b +size 510404157 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json new file mode 100644 index 0000000000000000000000000000000000000000..9a32a315de005a8e6606f2d0b64b117cc84d63c8 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873108.learnfair0431.3905918.0 b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873108.learnfair0431.3905918.0 new file mode 100644 index 0000000000000000000000000000000000000000..1b03a465904286b747dbbe97a18cc46be5cd5666 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873108.learnfair0431.3905918.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afa604c41d27d75ff45a5e8fef2cf2339ddb6f29ad1127b09c02eda02e2017ca +size 49380 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..325ca6483dc24d89d73db81a6a8773e32e22cc2d --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9871a175a7701be8a0237f4cba514aad13d664007361dd4753cd1e3293422545 +size 995653957 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..213e8aaa6c5059db7ddd27e5dc1a1c615c521b6a --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9312d596b21b3c2a9073e4763ab7b60c961fdc31925d88458c96f2fba7a544ae +size 510404157 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json new file mode 100644 index 0000000000000000000000000000000000000000..99b3b5de6277a85c9c30f8838c1dc2fa002a311f --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 2, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873101.learnfair0451.3484952.0 b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873101.learnfair0451.3484952.0 new file mode 100644 index 0000000000000000000000000000000000000000..4492e013c89e4525fb02673ce6868e36957ddf52 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873101.learnfair0451.3484952.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b8f4c0c9c652af535ed348f0fdfd2f82a8f8017a0342a7a51e45840b33000fa +size 49435 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3f6080fa4208bafb8b8efbc4605b905586a724c7 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "gpt2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50259 +} diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/global_step.txt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..9463411b62f21b7ed88bbe711c958b3b66153330 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/global_step.txt @@ -0,0 +1 @@ +2000 \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/optimizer.pt b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..44e0406d1c6961be73579ea3e9baa7da20be42c6 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d1dc19476afe61d0395b3820557156216e1491a2f9f9bb8c6568d86b3cf9d1d +size 995653957 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/pytorch_model.bin b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..010e4cca25afb8a2ad06176783a0adb20c34f432 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/checkpoint-2000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:408ff64cae8f9defe8fd7e1af16fd7878d059b7871e54d1ad77fa4707ddbf96e +size 510404157 diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json new file mode 100644 index 0000000000000000000000000000000000000000..8faa666c52f8b7538bab51545d6d2a886b8db744 --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": true, "seed": 42, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674872438.devfair0748.483334.0 b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674872438.devfair0748.483334.0 new file mode 100644 index 0000000000000000000000000000000000000000..af0422273e3ed73b3d087979e6ef18a39a81542c --- /dev/null +++ b/logs/sokoban/model:gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-42/events.out.tfevents.1674872438.devfair0748.483334.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:760f4f3f10a1a7da238369d906db17949f865473ed30b1c4d76d0d4e8439fbce +size 99020 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..be15b2ed9b3c86b42e3173b47610c0ea40e1507d --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/config.json @@ -0,0 +1,40 @@ +{ + "_name_or_path": "microsoft/CodeGPT-small-java-adaptedGPT2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "gradient_checkpointing": false, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50263 +} diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7d9de9579167e60e2b0641e0a9a11a2914c2fc80 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5114c278bc5b5e56bdb21906b6109a5f6c08f21e48da523bd76d202171adb857 +size 995678533 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b379d0bb80a978414835c124ed2a1edbbf1ff6e --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96aad474294e19fd33d57dfcd7f7302f2e643a267aaf621d3da35964f6f29c44 +size 510416445 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json new file mode 100644 index 0000000000000000000000000000000000000000..394eaf31e586ffde977c5a9519de91e5224a0491 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "java-gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 0, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873091.learnfair0402.3305517.0 b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873091.learnfair0402.3305517.0 new file mode 100644 index 0000000000000000000000000000000000000000..8415f172ab3e3cd0e0a63f7f69c16d5bb3e9078a --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-0/events.out.tfevents.1674873091.learnfair0402.3305517.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:373027df40704d4e8ee26ef54edc9d36d04ce31cc27ff4c7b2d778023588acfd +size 49369 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..be15b2ed9b3c86b42e3173b47610c0ea40e1507d --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/config.json @@ -0,0 +1,40 @@ +{ + "_name_or_path": "microsoft/CodeGPT-small-java-adaptedGPT2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "gradient_checkpointing": false, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50263 +} diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3b570e3aa4ad4a40fd5aaf746f19ba5354e60985 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e5655220c2b03f1ad52d8b79ad5112c27cb8667fcecc82aa391a853f8099594 +size 995678533 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..513c2476e062f3ced440477a590a499667405e8c --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae17def25bec87b84911cc12b1f601f18e299039021ac45a558b0b14cbd5bee3 +size 510416445 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7687a62f241f908e3b78de3038fbce235f791d57 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "java-gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 1, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873093.learnfair0438.3315824.0 b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873093.learnfair0438.3315824.0 new file mode 100644 index 0000000000000000000000000000000000000000..b0d9302276a08b653e0cf5779b248489916c0406 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-1/events.out.tfevents.1674873093.learnfair0438.3315824.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cafa2a85bd4f7bf0d3b32b169283735a0a3bdaa7b52a5785d76375719888e6b +size 49368 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..be15b2ed9b3c86b42e3173b47610c0ea40e1507d --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/config.json @@ -0,0 +1,40 @@ +{ + "_name_or_path": "microsoft/CodeGPT-small-java-adaptedGPT2", + "activation_function": "gelu_new", + "architectures": [ + "GPT2LMHeadModel" + ], + "attn_pdrop": 0.1, + "bos_token_id": 50256, + "embd_pdrop": 0.1, + "eos_token_id": 50256, + "gradient_checkpointing": false, + "initializer_range": 0.02, + "layer_norm_epsilon": 1e-05, + "model_type": "gpt2", + "n_ctx": 1024, + "n_embd": 768, + "n_head": 12, + "n_inner": null, + "n_layer": 12, + "n_positions": 1024, + "reorder_and_upcast_attn": false, + "resid_pdrop": 0.1, + "scale_attn_by_inverse_layer_idx": false, + "scale_attn_weights": true, + "summary_activation": null, + "summary_first_dropout": 0.1, + "summary_proj_to_labels": true, + "summary_type": "cls_index", + "summary_use_proj": true, + "task_specific_params": { + "text-generation": { + "do_sample": true, + "max_length": 50 + } + }, + "torch_dtype": "float32", + "transformers_version": "4.25.1", + "use_cache": true, + "vocab_size": 50263 +} diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt new file mode 100644 index 0000000000000000000000000000000000000000..e37d32abba426c06b752a5e53f48c595c84e9270 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/global_step.txt @@ -0,0 +1 @@ +1000 \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..bc7f941b7e8f32c12155cad852cc52b30b2f818f --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae2390b9731c835f0c4885cfb524f0b342e8e5216ffd907e761ce46ad1c8e547 +size 995678533 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin new file mode 100644 index 0000000000000000000000000000000000000000..0860534b19ab31be38c439848b1bba178ec81e8b --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/checkpoint-1000/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67de140b6509e8b6d5d314036631dabb8aeecb30b329d169d107d3878620a81b +size 510416445 diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json new file mode 100644 index 0000000000000000000000000000000000000000..13fe9c64861da0503ab64dbd2e406efbaa6631d7 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/config.json @@ -0,0 +1 @@ +{"game": "sokoban", "level_key": "level", "annotation_keys": ["solution_len"], "num_annotation_buckets": null, "holdout_solution_lens": null, "chunk_size": 128, "model": "java-gpt2", "warmup_proportion": 0.0002, "weight_decay": 0.01, "max_grad_norm": 1, "learning_rate": 0.0001, "exp_name": "", "overwrite": false, "seed": 2, "batch_size": 16, "epochs": 20, "save_freq": 1000, "eval_freq": 1000, "no_log": false, "num_eval_samples": 20, "gen_freq": 500, "gen_len": 128, "gen_temp": 1.0, "gen_beams": 5, "gen_top_k": 50, "gen_top_p": 1.0, "gen_typical_p": 1.0, "sample_contexts": false} \ No newline at end of file diff --git a/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873092.learnfair0488.331314.0 b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873092.learnfair0488.331314.0 new file mode 100644 index 0000000000000000000000000000000000000000..d2c8de9944c4109e7f1df9fbb35177f43fc56552 --- /dev/null +++ b/logs/sokoban/model:java-gpt2/level_key:level/annotation_keys:['solution_len']/num_annotation_buckets:None/holdouts:None/chunk_size-128_lr-0.0001/seed-2/events.out.tfevents.1674873092.learnfair0488.331314.0 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f19cca5fb1cbdf931e8faeca5faeeacdbb4aadff9366bf40f1203b71d8937f2 +size 49378