joshuaclymer commited on
Commit
35936b6
·
1 Parent(s): c47a54e

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - quant_method: bitsandbytes
9
+ - load_in_8bit: False
10
+ - load_in_4bit: True
11
+ - llm_int8_threshold: 6.0
12
+ - llm_int8_skip_modules: None
13
+ - llm_int8_enable_fp32_cpu_offload: False
14
+ - llm_int8_has_fp16_weight: False
15
+ - bnb_4bit_quant_type: nf4
16
+ - bnb_4bit_use_double_quant: True
17
+ - bnb_4bit_compute_dtype: float16
18
+ ### Framework versions
19
+
20
+
21
+ - PEFT 0.5.0
adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "models/llama-13b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "v_proj"
19
+ ],
20
+ "task_type": "SEQ_CLS"
21
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:339bf4746ef195444df4158a994d78e91cd5f9c1a452dc8c9db63eed279ca8fa
3
+ size 209811921
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "__type": "AddedToken",
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ "clean_up_tokenization_spaces": false,
11
+ "eos_token": {
12
+ "__type": "AddedToken",
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "model_max_length": 2048,
20
+ "pad_token": null,
21
+ "sp_model_kwargs": {},
22
+ "tokenizer_class": "LlamaTokenizer",
23
+ "unk_token": {
24
+ "__type": "AddedToken",
25
+ "content": "<unk>",
26
+ "lstrip": false,
27
+ "normalized": true,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ "use_default_system_prompt": true
32
+ }
train_args.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "initial_model_dir": "models/llama-13b",
3
+ "distribution_id": "code_easy",
4
+ "date_trained": "10/10/2023 11:15:41",
5
+ "output_dir": "models/classify_lora/llama-13b-code_easy",
6
+ "overwrite_output_dir": false,
7
+ "do_train": false,
8
+ "do_eval": true,
9
+ "do_predict": false,
10
+ "evaluation_strategy": "steps",
11
+ "prediction_loss_only": false,
12
+ "per_device_train_batch_size": 16,
13
+ "per_device_eval_batch_size": 16,
14
+ "per_gpu_train_batch_size": null,
15
+ "per_gpu_eval_batch_size": null,
16
+ "gradient_accumulation_steps": 1,
17
+ "eval_accumulation_steps": null,
18
+ "eval_delay": 0,
19
+ "learning_rate": 0.0002,
20
+ "weight_decay": 0.0,
21
+ "adam_beta1": 0.9,
22
+ "adam_beta2": 0.999,
23
+ "adam_epsilon": 1e-08,
24
+ "max_grad_norm": 0.3,
25
+ "num_train_epochs": 3,
26
+ "max_steps": 150,
27
+ "lr_scheduler_type": "constant",
28
+ "warmup_ratio": 0.03,
29
+ "warmup_steps": 0,
30
+ "log_level": "passive",
31
+ "log_level_replica": "warning",
32
+ "log_on_each_node": true,
33
+ "logging_dir": "models/classify_lora/llama-13b-code_easy/runs/Oct10_09-53-32_compute-permanent-node-978",
34
+ "logging_strategy": "steps",
35
+ "logging_first_step": false,
36
+ "logging_steps": 1,
37
+ "logging_nan_inf_filter": true,
38
+ "save_strategy": "steps",
39
+ "save_steps": 25,
40
+ "save_total_limit": 0,
41
+ "save_safetensors": false,
42
+ "save_on_each_node": false,
43
+ "no_cuda": false,
44
+ "use_cpu": false,
45
+ "use_mps_device": false,
46
+ "seed": 42,
47
+ "data_seed": null,
48
+ "jit_mode_eval": false,
49
+ "use_ipex": false,
50
+ "bf16": false,
51
+ "fp16": false,
52
+ "fp16_opt_level": "O1",
53
+ "half_precision_backend": "auto",
54
+ "bf16_full_eval": false,
55
+ "fp16_full_eval": false,
56
+ "tf32": null,
57
+ "local_rank": 1,
58
+ "ddp_backend": null,
59
+ "tpu_num_cores": null,
60
+ "tpu_metrics_debug": false,
61
+ "debug": [],
62
+ "dataloader_drop_last": false,
63
+ "eval_steps": 25,
64
+ "dataloader_num_workers": 0,
65
+ "past_index": -1,
66
+ "run_name": "train|models-classify_lora-llama-13b-code_easy",
67
+ "disable_tqdm": false,
68
+ "remove_unused_columns": false,
69
+ "label_names": null,
70
+ "load_best_model_at_end": false,
71
+ "metric_for_best_model": "eval_code_easy_score",
72
+ "greater_is_better": true,
73
+ "ignore_data_skip": false,
74
+ "sharded_ddp": [],
75
+ "fsdp": [],
76
+ "fsdp_min_num_params": 0,
77
+ "fsdp_config": {
78
+ "min_num_params": 0,
79
+ "xla": false,
80
+ "xla_fsdp_grad_ckpt": false
81
+ },
82
+ "fsdp_transformer_layer_cls_to_wrap": null,
83
+ "deepspeed": "configs/ds_zero_1.json",
84
+ "label_smoothing_factor": 0.0,
85
+ "optim": "paged_adamw_32bit",
86
+ "optim_args": null,
87
+ "adafactor": false,
88
+ "group_by_length": false,
89
+ "length_column_name": "length",
90
+ "report_to": [
91
+ "wandb"
92
+ ],
93
+ "ddp_find_unused_parameters": false,
94
+ "ddp_bucket_cap_mb": null,
95
+ "ddp_broadcast_buffers": null,
96
+ "dataloader_pin_memory": true,
97
+ "skip_memory_metrics": true,
98
+ "use_legacy_prediction_loop": false,
99
+ "push_to_hub": false,
100
+ "resume_from_checkpoint": null,
101
+ "hub_model_id": null,
102
+ "hub_strategy": "every_save",
103
+ "hub_token": null,
104
+ "hub_private_repo": false,
105
+ "hub_always_push": false,
106
+ "gradient_checkpointing": false,
107
+ "include_inputs_for_metrics": false,
108
+ "fp16_backend": "auto",
109
+ "push_to_hub_model_id": null,
110
+ "push_to_hub_organization": null,
111
+ "push_to_hub_token": null,
112
+ "_n_gpu": 1,
113
+ "mp_parameters": "",
114
+ "auto_find_batch_size": false,
115
+ "full_determinism": false,
116
+ "torchdynamo": null,
117
+ "ray_scope": "last",
118
+ "ddp_timeout": 1800,
119
+ "torch_compile": false,
120
+ "torch_compile_backend": null,
121
+ "torch_compile_mode": null,
122
+ "dispatch_batches": null
123
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51570299200d286eb232ae30708923f074d49b1c3e4b74c8a514ff4cd334b077
3
+ size 5115
training_logs.json ADDED
@@ -0,0 +1,2279 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "loss": 0.7696,
4
+ "learning_rate": 0.0002,
5
+ "epoch": 0.02,
6
+ "step": 1
7
+ },
8
+ {
9
+ "loss": 0.6496,
10
+ "learning_rate": 0.0002,
11
+ "epoch": 0.04,
12
+ "step": 2
13
+ },
14
+ {
15
+ "loss": 0.6886,
16
+ "learning_rate": 0.0002,
17
+ "epoch": 0.05,
18
+ "step": 3
19
+ },
20
+ {
21
+ "loss": 0.6262,
22
+ "learning_rate": 0.0002,
23
+ "epoch": 0.07,
24
+ "step": 4
25
+ },
26
+ {
27
+ "loss": 0.5216,
28
+ "learning_rate": 0.0002,
29
+ "epoch": 0.09,
30
+ "step": 5
31
+ },
32
+ {
33
+ "loss": 0.5246,
34
+ "learning_rate": 0.0002,
35
+ "epoch": 0.11,
36
+ "step": 6
37
+ },
38
+ {
39
+ "loss": 0.4545,
40
+ "learning_rate": 0.0002,
41
+ "epoch": 0.12,
42
+ "step": 7
43
+ },
44
+ {
45
+ "loss": 0.4518,
46
+ "learning_rate": 0.0002,
47
+ "epoch": 0.14,
48
+ "step": 8
49
+ },
50
+ {
51
+ "loss": 0.3915,
52
+ "learning_rate": 0.0002,
53
+ "epoch": 0.16,
54
+ "step": 9
55
+ },
56
+ {
57
+ "loss": 0.4736,
58
+ "learning_rate": 0.0002,
59
+ "epoch": 0.18,
60
+ "step": 10
61
+ },
62
+ {
63
+ "loss": 0.4215,
64
+ "learning_rate": 0.0002,
65
+ "epoch": 0.19,
66
+ "step": 11
67
+ },
68
+ {
69
+ "loss": 0.3712,
70
+ "learning_rate": 0.0002,
71
+ "epoch": 0.21,
72
+ "step": 12
73
+ },
74
+ {
75
+ "loss": 0.3054,
76
+ "learning_rate": 0.0002,
77
+ "epoch": 0.23,
78
+ "step": 13
79
+ },
80
+ {
81
+ "loss": 0.2136,
82
+ "learning_rate": 0.0002,
83
+ "epoch": 0.25,
84
+ "step": 14
85
+ },
86
+ {
87
+ "loss": 0.1414,
88
+ "learning_rate": 0.0002,
89
+ "epoch": 0.26,
90
+ "step": 15
91
+ },
92
+ {
93
+ "loss": 0.215,
94
+ "learning_rate": 0.0002,
95
+ "epoch": 0.28,
96
+ "step": 16
97
+ },
98
+ {
99
+ "loss": 0.3062,
100
+ "learning_rate": 0.0002,
101
+ "epoch": 0.3,
102
+ "step": 17
103
+ },
104
+ {
105
+ "loss": 0.3171,
106
+ "learning_rate": 0.0002,
107
+ "epoch": 0.32,
108
+ "step": 18
109
+ },
110
+ {
111
+ "loss": 0.2465,
112
+ "learning_rate": 0.0002,
113
+ "epoch": 0.33,
114
+ "step": 19
115
+ },
116
+ {
117
+ "loss": 0.2049,
118
+ "learning_rate": 0.0002,
119
+ "epoch": 0.35,
120
+ "step": 20
121
+ },
122
+ {
123
+ "loss": 0.2947,
124
+ "learning_rate": 0.0002,
125
+ "epoch": 0.37,
126
+ "step": 21
127
+ },
128
+ {
129
+ "loss": 0.1174,
130
+ "learning_rate": 0.0002,
131
+ "epoch": 0.39,
132
+ "step": 22
133
+ },
134
+ {
135
+ "loss": 0.118,
136
+ "learning_rate": 0.0002,
137
+ "epoch": 0.4,
138
+ "step": 23
139
+ },
140
+ {
141
+ "loss": 0.1616,
142
+ "learning_rate": 0.0002,
143
+ "epoch": 0.42,
144
+ "step": 24
145
+ },
146
+ {
147
+ "loss": 0.1731,
148
+ "learning_rate": 0.0002,
149
+ "epoch": 0.44,
150
+ "step": 25
151
+ },
152
+ {
153
+ "eval_code_easy_loss": 0.1410333812236786,
154
+ "eval_code_easy_score": -0.04131929576396942,
155
+ "eval_code_easy_brier_score": 0.04131929576396942,
156
+ "eval_code_easy_average_probability": 0.8963662981987,
157
+ "eval_code_easy_accuracy": 0.94,
158
+ "eval_code_easy_probabilities": [
159
+ 0.9897421598434448,
160
+ 0.9846920371055603,
161
+ 0.9921024441719055,
162
+ 0.9999797344207764,
163
+ 0.9999569654464722,
164
+ 0.9998584985733032,
165
+ 0.4596724212169647,
166
+ 0.9999803304672241,
167
+ 0.9999934434890747,
168
+ 0.6896206736564636,
169
+ 0.7018794417381287,
170
+ 0.8044198751449585,
171
+ 0.9061204791069031,
172
+ 0.9999116659164429,
173
+ 0.9998483657836914,
174
+ 0.9975529313087463,
175
+ 0.9999861717224121,
176
+ 0.9999945163726807,
177
+ 0.9997785687446594,
178
+ 0.9217097163200378,
179
+ 0.549553632736206,
180
+ 0.9999867677688599,
181
+ 0.9999836683273315,
182
+ 0.9999502897262573,
183
+ 0.9999196529388428,
184
+ 0.999788224697113,
185
+ 0.9994627833366394,
186
+ 0.8858318328857422,
187
+ 0.8768848180770874,
188
+ 0.9063717126846313,
189
+ 0.6450658440589905,
190
+ 0.718223512172699,
191
+ 0.6423218846321106,
192
+ 0.6255131363868713,
193
+ 0.4555876851081848,
194
+ 0.5870699286460876,
195
+ 0.7827481031417847,
196
+ 0.9729915857315063,
197
+ 0.7549998164176941,
198
+ 0.9984861016273499,
199
+ 0.9989733695983887,
200
+ 0.9916539192199707,
201
+ 0.9056527018547058,
202
+ 0.8946619629859924,
203
+ 0.8330284357070923,
204
+ 0.9995453953742981,
205
+ 0.5416378378868103,
206
+ 0.9464252591133118,
207
+ 0.972342848777771,
208
+ 0.9960533380508423,
209
+ 0.988210916519165,
210
+ 0.4754994809627533,
211
+ 0.9999995231628418,
212
+ 0.9999991655349731,
213
+ 0.9999830722808838,
214
+ 0.9999951124191284,
215
+ 0.9999954700469971,
216
+ 0.9999794960021973,
217
+ 0.3308394253253937,
218
+ 0.9999799728393555,
219
+ 0.9992032647132874,
220
+ 0.9999996423721313,
221
+ 0.9999727010726929,
222
+ 0.9999815225601196,
223
+ 0.9999794960021973,
224
+ 0.9999793767929077,
225
+ 0.988341748714447,
226
+ 0.99339359998703,
227
+ 0.9851863980293274,
228
+ 0.631180465221405,
229
+ 0.2889827787876129,
230
+ 0.2886982560157776,
231
+ 0.9839469194412231,
232
+ 0.9860828518867493,
233
+ 0.9651493430137634,
234
+ 0.7786540389060974,
235
+ 0.8290436863899231,
236
+ 0.9590767621994019,
237
+ 0.9999992847442627,
238
+ 0.9999868869781494,
239
+ 0.9999861717224121,
240
+ 0.9500917196273804,
241
+ 0.9666972160339355,
242
+ 0.8169301748275757,
243
+ 0.9984229803085327,
244
+ 0.9998488426208496,
245
+ 0.9996005892753601,
246
+ 0.9686463475227356,
247
+ 0.9352719783782959,
248
+ 0.9167892336845398,
249
+ 0.9999650716781616,
250
+ 0.8374965190887451,
251
+ 0.9999854564666748,
252
+ 0.9999997615814209,
253
+ 0.9999997615814209,
254
+ 0.9999992847442627,
255
+ 0.9999997615814209,
256
+ 0.9999991655349731,
257
+ 0.9999997615814209,
258
+ 0.848063588142395
259
+ ],
260
+ "eval_code_easy_runtime": 31.1515,
261
+ "eval_code_easy_samples_per_second": 3.21,
262
+ "eval_code_easy_steps_per_second": 0.128,
263
+ "epoch": 0.44,
264
+ "step": 25
265
+ },
266
+ {
267
+ "eval_code_hard_loss": 0.44707271456718445,
268
+ "eval_code_hard_score": -0.14680194854736328,
269
+ "eval_code_hard_brier_score": 0.14680194854736328,
270
+ "eval_code_hard_average_probability": 0.6711300015449524,
271
+ "eval_code_hard_accuracy": 0.83,
272
+ "eval_code_hard_probabilities": [
273
+ 0.9999589920043945,
274
+ 0.47546130418777466,
275
+ 0.5713449716567993,
276
+ 0.9994718432426453,
277
+ 0.9991668462753296,
278
+ 0.9993312358856201,
279
+ 0.4067853093147278,
280
+ 0.40341079235076904,
281
+ 0.41259637475013733,
282
+ 0.4262898862361908,
283
+ 0.48981818556785583,
284
+ 0.6610403060913086,
285
+ 0.9994863271713257,
286
+ 0.9982219338417053,
287
+ 0.99834144115448,
288
+ 0.7558809518814087,
289
+ 0.6139836311340332,
290
+ 0.6215753555297852,
291
+ 0.9995869994163513,
292
+ 0.9992690682411194,
293
+ 0.9992164373397827,
294
+ 0.8080712556838989,
295
+ 0.8023390769958496,
296
+ 0.7505528330802917,
297
+ 0.5108107328414917,
298
+ 0.5139395594596863,
299
+ 0.4423873722553253,
300
+ 0.7270267605781555,
301
+ 0.7878870368003845,
302
+ 0.7860028743743896,
303
+ 0.6687493920326233,
304
+ 0.7331302165985107,
305
+ 0.6188584566116333,
306
+ 0.5269416570663452,
307
+ 0.4910818338394165,
308
+ 0.4837784469127655,
309
+ 0.5938860774040222,
310
+ 0.5307540893554688,
311
+ 0.6091882586479187,
312
+ 0.5714852213859558,
313
+ 0.5575187802314758,
314
+ 0.636616587638855,
315
+ 0.5653713941574097,
316
+ 0.5363364219665527,
317
+ 0.5516955256462097,
318
+ 0.7773237228393555,
319
+ 0.5234290361404419,
320
+ 0.5288711786270142,
321
+ 0.6337161064147949,
322
+ 0.6571640968322754,
323
+ 0.6706582903862,
324
+ 0.45844873785972595,
325
+ 0.7167773842811584,
326
+ 0.605722188949585,
327
+ 0.9969874024391174,
328
+ 0.9969995617866516,
329
+ 0.9975918531417847,
330
+ 0.3677201271057129,
331
+ 0.5539365410804749,
332
+ 0.7323344349861145,
333
+ 0.9991826415061951,
334
+ 0.9991852641105652,
335
+ 0.9992254972457886,
336
+ 0.9554563760757446,
337
+ 0.9320324659347534,
338
+ 0.9908183217048645,
339
+ 0.5780607461929321,
340
+ 0.9173429608345032,
341
+ 0.573617160320282,
342
+ 0.710330605506897,
343
+ 0.610026478767395,
344
+ 0.5195437073707581,
345
+ 0.8946442008018494,
346
+ 0.925991415977478,
347
+ 0.938264787197113,
348
+ 0.5608754754066467,
349
+ 0.5026195049285889,
350
+ 0.4715512990951538,
351
+ 0.5032576322555542,
352
+ 0.5153825879096985,
353
+ 0.5001385807991028,
354
+ 0.6857117414474487,
355
+ 0.6296961307525635,
356
+ 0.58381187915802,
357
+ 0.5933532118797302,
358
+ 0.5141075253486633,
359
+ 0.5213065147399902,
360
+ 0.4824189841747284,
361
+ 0.5641835331916809,
362
+ 0.5198292136192322,
363
+ 0.5848154425621033,
364
+ 0.6065869927406311,
365
+ 0.5906622409820557,
366
+ 0.39757779240608215,
367
+ 0.37057143449783325,
368
+ 0.33965691924095154,
369
+ 0.638043999671936,
370
+ 0.75816410779953,
371
+ 0.8119453191757202,
372
+ 0.4706781804561615
373
+ ],
374
+ "eval_code_hard_runtime": 110.7399,
375
+ "eval_code_hard_samples_per_second": 0.903,
376
+ "eval_code_hard_steps_per_second": 0.036,
377
+ "epoch": 0.44,
378
+ "step": 25
379
+ },
380
+ {
381
+ "loss": 0.1187,
382
+ "learning_rate": 0.0002,
383
+ "epoch": 0.46,
384
+ "step": 26
385
+ },
386
+ {
387
+ "loss": 0.1277,
388
+ "learning_rate": 0.0002,
389
+ "epoch": 0.47,
390
+ "step": 27
391
+ },
392
+ {
393
+ "loss": 0.0581,
394
+ "learning_rate": 0.0002,
395
+ "epoch": 0.49,
396
+ "step": 28
397
+ },
398
+ {
399
+ "loss": 0.1622,
400
+ "learning_rate": 0.0002,
401
+ "epoch": 0.51,
402
+ "step": 29
403
+ },
404
+ {
405
+ "loss": 0.0558,
406
+ "learning_rate": 0.0002,
407
+ "epoch": 0.53,
408
+ "step": 30
409
+ },
410
+ {
411
+ "loss": 0.0873,
412
+ "learning_rate": 0.0002,
413
+ "epoch": 0.54,
414
+ "step": 31
415
+ },
416
+ {
417
+ "loss": 0.1237,
418
+ "learning_rate": 0.0002,
419
+ "epoch": 0.56,
420
+ "step": 32
421
+ },
422
+ {
423
+ "loss": 0.1127,
424
+ "learning_rate": 0.0002,
425
+ "epoch": 0.58,
426
+ "step": 33
427
+ },
428
+ {
429
+ "loss": 0.0673,
430
+ "learning_rate": 0.0002,
431
+ "epoch": 0.6,
432
+ "step": 34
433
+ },
434
+ {
435
+ "loss": 0.0628,
436
+ "learning_rate": 0.0002,
437
+ "epoch": 0.61,
438
+ "step": 35
439
+ },
440
+ {
441
+ "loss": 0.1647,
442
+ "learning_rate": 0.0002,
443
+ "epoch": 0.63,
444
+ "step": 36
445
+ },
446
+ {
447
+ "loss": 0.0389,
448
+ "learning_rate": 0.0002,
449
+ "epoch": 0.65,
450
+ "step": 37
451
+ },
452
+ {
453
+ "loss": 0.0162,
454
+ "learning_rate": 0.0002,
455
+ "epoch": 0.67,
456
+ "step": 38
457
+ },
458
+ {
459
+ "loss": 0.1205,
460
+ "learning_rate": 0.0002,
461
+ "epoch": 0.68,
462
+ "step": 39
463
+ },
464
+ {
465
+ "loss": 0.2798,
466
+ "learning_rate": 0.0002,
467
+ "epoch": 0.7,
468
+ "step": 40
469
+ },
470
+ {
471
+ "loss": 0.0202,
472
+ "learning_rate": 0.0002,
473
+ "epoch": 0.72,
474
+ "step": 41
475
+ },
476
+ {
477
+ "loss": 0.0563,
478
+ "learning_rate": 0.0002,
479
+ "epoch": 0.74,
480
+ "step": 42
481
+ },
482
+ {
483
+ "loss": 0.0959,
484
+ "learning_rate": 0.0002,
485
+ "epoch": 0.75,
486
+ "step": 43
487
+ },
488
+ {
489
+ "loss": 0.033,
490
+ "learning_rate": 0.0002,
491
+ "epoch": 0.77,
492
+ "step": 44
493
+ },
494
+ {
495
+ "loss": 0.0377,
496
+ "learning_rate": 0.0002,
497
+ "epoch": 0.79,
498
+ "step": 45
499
+ },
500
+ {
501
+ "loss": 0.1312,
502
+ "learning_rate": 0.0002,
503
+ "epoch": 0.81,
504
+ "step": 46
505
+ },
506
+ {
507
+ "loss": 0.1171,
508
+ "learning_rate": 0.0002,
509
+ "epoch": 0.82,
510
+ "step": 47
511
+ },
512
+ {
513
+ "loss": 0.0586,
514
+ "learning_rate": 0.0002,
515
+ "epoch": 0.84,
516
+ "step": 48
517
+ },
518
+ {
519
+ "loss": 0.0373,
520
+ "learning_rate": 0.0002,
521
+ "epoch": 0.86,
522
+ "step": 49
523
+ },
524
+ {
525
+ "loss": 0.1418,
526
+ "learning_rate": 0.0002,
527
+ "epoch": 0.88,
528
+ "step": 50
529
+ },
530
+ {
531
+ "eval_code_easy_loss": 0.09489491581916809,
532
+ "eval_code_easy_score": -0.030822403728961945,
533
+ "eval_code_easy_brier_score": 0.030822403728961945,
534
+ "eval_code_easy_average_probability": 0.9424028992652893,
535
+ "eval_code_easy_accuracy": 0.95,
536
+ "eval_code_easy_probabilities": [
537
+ 0.9999914169311523,
538
+ 1.0,
539
+ 1.0,
540
+ 0.9999921321868896,
541
+ 0.9999855756759644,
542
+ 0.9999556541442871,
543
+ 0.4764171540737152,
544
+ 1.0,
545
+ 1.0,
546
+ 0.9999983310699463,
547
+ 0.9999955892562866,
548
+ 0.9999998807907104,
549
+ 0.9997754693031311,
550
+ 0.9999978542327881,
551
+ 0.9999951124191284,
552
+ 0.9944272041320801,
553
+ 1.0,
554
+ 1.0,
555
+ 1.0,
556
+ 0.9332470893859863,
557
+ 0.686340868473053,
558
+ 0.9999998807907104,
559
+ 0.9999998807907104,
560
+ 0.9999996423721313,
561
+ 1.0,
562
+ 0.9999998807907104,
563
+ 0.9999986886978149,
564
+ 0.9998589754104614,
565
+ 0.9998730421066284,
566
+ 0.9999005794525146,
567
+ 0.7265980839729309,
568
+ 0.8790691494941711,
569
+ 0.8301546573638916,
570
+ 0.9983548521995544,
571
+ 0.9948050379753113,
572
+ 0.9825286865234375,
573
+ 0.962716817855835,
574
+ 0.9991486072540283,
575
+ 0.9315124750137329,
576
+ 0.9939038157463074,
577
+ 0.9942259192466736,
578
+ 0.9889645576477051,
579
+ 0.9977827668190002,
580
+ 0.9999998807907104,
581
+ 0.9890486001968384,
582
+ 0.9997009038925171,
583
+ 0.20336638391017914,
584
+ 0.9882842302322388,
585
+ 0.9745821952819824,
586
+ 0.9941431879997253,
587
+ 0.9512994289398193,
588
+ 0.7942199110984802,
589
+ 1.0,
590
+ 1.0,
591
+ 1.0,
592
+ 1.0,
593
+ 1.0,
594
+ 0.999992847442627,
595
+ 0.254838228225708,
596
+ 0.9999922513961792,
597
+ 1.0,
598
+ 1.0,
599
+ 0.9999960660934448,
600
+ 1.0,
601
+ 1.0,
602
+ 1.0,
603
+ 1.0,
604
+ 1.0,
605
+ 1.0,
606
+ 0.9282342791557312,
607
+ 0.25453829765319824,
608
+ 0.1509198695421219,
609
+ 0.9998898506164551,
610
+ 0.9999701976776123,
611
+ 0.999139666557312,
612
+ 0.8278753161430359,
613
+ 0.8814498782157898,
614
+ 0.9445298910140991,
615
+ 1.0,
616
+ 0.9999996423721313,
617
+ 1.0,
618
+ 0.9991635084152222,
619
+ 1.0,
620
+ 0.9999971389770508,
621
+ 0.955471932888031,
622
+ 0.9967992305755615,
623
+ 0.9889751672744751,
624
+ 0.9923612475395203,
625
+ 0.9318004250526428,
626
+ 0.9118591547012329,
627
+ 1.0,
628
+ 0.95877605676651,
629
+ 0.9997400641441345,
630
+ 1.0,
631
+ 1.0,
632
+ 1.0,
633
+ 1.0,
634
+ 1.0,
635
+ 1.0,
636
+ 0.9998133778572083
637
+ ],
638
+ "eval_code_easy_runtime": 31.1754,
639
+ "eval_code_easy_samples_per_second": 3.208,
640
+ "eval_code_easy_steps_per_second": 0.128,
641
+ "epoch": 0.88,
642
+ "step": 50
643
+ },
644
+ {
645
+ "eval_code_hard_loss": 0.48955029249191284,
646
+ "eval_code_hard_score": -0.16734105348587036,
647
+ "eval_code_hard_brier_score": 0.16734105348587036,
648
+ "eval_code_hard_average_probability": 0.6507197022438049,
649
+ "eval_code_hard_accuracy": 0.74,
650
+ "eval_code_hard_probabilities": [
651
+ 0.998719334602356,
652
+ 0.4495692551136017,
653
+ 0.560586154460907,
654
+ 0.9998689889907837,
655
+ 0.9994572997093201,
656
+ 0.999840259552002,
657
+ 0.40500903129577637,
658
+ 0.4450298845767975,
659
+ 0.39399078488349915,
660
+ 0.48165562748908997,
661
+ 0.4001981019973755,
662
+ 0.6641137599945068,
663
+ 0.999977707862854,
664
+ 0.9998791217803955,
665
+ 0.9998584985733032,
666
+ 0.8799082636833191,
667
+ 0.766524612903595,
668
+ 0.775595486164093,
669
+ 0.9999852180480957,
670
+ 0.9999836683273315,
671
+ 0.9999798536300659,
672
+ 0.8477662205696106,
673
+ 0.7681900262832642,
674
+ 0.7254011631011963,
675
+ 0.4458688199520111,
676
+ 0.4710305333137512,
677
+ 0.3777993619441986,
678
+ 0.09769346565008163,
679
+ 0.4915977418422699,
680
+ 0.5548089146614075,
681
+ 0.6629347801208496,
682
+ 0.7435128092765808,
683
+ 0.5462533831596375,
684
+ 0.6541544198989868,
685
+ 0.4705035388469696,
686
+ 0.5168669819831848,
687
+ 0.42897677421569824,
688
+ 0.3096539080142975,
689
+ 0.4461301267147064,
690
+ 0.461626797914505,
691
+ 0.46588248014450073,
692
+ 0.5764793157577515,
693
+ 0.5880526304244995,
694
+ 0.5188709497451782,
695
+ 0.5498936772346497,
696
+ 0.8593714237213135,
697
+ 0.5482467412948608,
698
+ 0.5750262141227722,
699
+ 0.67604660987854,
700
+ 0.8179431557655334,
701
+ 0.8246544003486633,
702
+ 0.4119476079940796,
703
+ 0.7206862568855286,
704
+ 0.6743555068969727,
705
+ 0.9999415874481201,
706
+ 0.9999454021453857,
707
+ 0.9999594688415527,
708
+ 0.470983624458313,
709
+ 0.5161882042884827,
710
+ 0.5438310503959656,
711
+ 0.9999815225601196,
712
+ 0.9999749660491943,
713
+ 0.9999841451644897,
714
+ 0.9674224257469177,
715
+ 0.9396224617958069,
716
+ 0.9942786693572998,
717
+ 0.6115734577178955,
718
+ 0.9555063247680664,
719
+ 0.6700289249420166,
720
+ 0.8057522177696228,
721
+ 0.6384589672088623,
722
+ 0.5619309544563293,
723
+ 0.6012542843818665,
724
+ 0.6818827986717224,
725
+ 0.709420382976532,
726
+ 0.6579272747039795,
727
+ 0.42321380972862244,
728
+ 0.44099465012550354,
729
+ 0.5047284364700317,
730
+ 0.506719172000885,
731
+ 0.5062504410743713,
732
+ 0.6765443682670593,
733
+ 0.7465528249740601,
734
+ 0.6119047999382019,
735
+ 0.5429056882858276,
736
+ 0.5111532807350159,
737
+ 0.5241301655769348,
738
+ 0.4921115040779114,
739
+ 0.6113234162330627,
740
+ 0.5521866679191589,
741
+ 0.5868935585021973,
742
+ 0.6085503697395325,
743
+ 0.5654861927032471,
744
+ 0.3610905706882477,
745
+ 0.3744204640388489,
746
+ 0.31809136271476746,
747
+ 0.5574522018432617,
748
+ 0.641338586807251,
749
+ 0.48631271719932556,
750
+ 0.5477950572967529
751
+ ],
752
+ "eval_code_hard_runtime": 110.6958,
753
+ "eval_code_hard_samples_per_second": 0.903,
754
+ "eval_code_hard_steps_per_second": 0.036,
755
+ "epoch": 0.88,
756
+ "step": 50
757
+ },
758
+ {
759
+ "loss": 0.0819,
760
+ "learning_rate": 0.0002,
761
+ "epoch": 0.89,
762
+ "step": 51
763
+ },
764
+ {
765
+ "loss": 0.0037,
766
+ "learning_rate": 0.0002,
767
+ "epoch": 0.91,
768
+ "step": 52
769
+ },
770
+ {
771
+ "loss": 0.1525,
772
+ "learning_rate": 0.0002,
773
+ "epoch": 0.93,
774
+ "step": 53
775
+ },
776
+ {
777
+ "loss": 0.2635,
778
+ "learning_rate": 0.0002,
779
+ "epoch": 0.95,
780
+ "step": 54
781
+ },
782
+ {
783
+ "loss": 0.1162,
784
+ "learning_rate": 0.0002,
785
+ "epoch": 0.96,
786
+ "step": 55
787
+ },
788
+ {
789
+ "loss": 0.0352,
790
+ "learning_rate": 0.0002,
791
+ "epoch": 0.98,
792
+ "step": 56
793
+ },
794
+ {
795
+ "loss": 0.1891,
796
+ "learning_rate": 0.0002,
797
+ "epoch": 1.0,
798
+ "step": 57
799
+ },
800
+ {
801
+ "loss": 0.0453,
802
+ "learning_rate": 0.0002,
803
+ "epoch": 1.02,
804
+ "step": 58
805
+ },
806
+ {
807
+ "loss": 0.0562,
808
+ "learning_rate": 0.0002,
809
+ "epoch": 1.04,
810
+ "step": 59
811
+ },
812
+ {
813
+ "loss": 0.0006,
814
+ "learning_rate": 0.0002,
815
+ "epoch": 1.05,
816
+ "step": 60
817
+ },
818
+ {
819
+ "loss": 0.0445,
820
+ "learning_rate": 0.0002,
821
+ "epoch": 1.07,
822
+ "step": 61
823
+ },
824
+ {
825
+ "loss": 0.0125,
826
+ "learning_rate": 0.0002,
827
+ "epoch": 1.09,
828
+ "step": 62
829
+ },
830
+ {
831
+ "loss": 0.0378,
832
+ "learning_rate": 0.0002,
833
+ "epoch": 1.11,
834
+ "step": 63
835
+ },
836
+ {
837
+ "loss": 0.0367,
838
+ "learning_rate": 0.0002,
839
+ "epoch": 1.12,
840
+ "step": 64
841
+ },
842
+ {
843
+ "loss": 0.0261,
844
+ "learning_rate": 0.0002,
845
+ "epoch": 1.14,
846
+ "step": 65
847
+ },
848
+ {
849
+ "loss": 0.0022,
850
+ "learning_rate": 0.0002,
851
+ "epoch": 1.16,
852
+ "step": 66
853
+ },
854
+ {
855
+ "loss": 0.0151,
856
+ "learning_rate": 0.0002,
857
+ "epoch": 1.18,
858
+ "step": 67
859
+ },
860
+ {
861
+ "loss": 0.0262,
862
+ "learning_rate": 0.0002,
863
+ "epoch": 1.19,
864
+ "step": 68
865
+ },
866
+ {
867
+ "loss": 0.0025,
868
+ "learning_rate": 0.0002,
869
+ "epoch": 1.21,
870
+ "step": 69
871
+ },
872
+ {
873
+ "loss": 0.009,
874
+ "learning_rate": 0.0002,
875
+ "epoch": 1.23,
876
+ "step": 70
877
+ },
878
+ {
879
+ "loss": 0.0233,
880
+ "learning_rate": 0.0002,
881
+ "epoch": 1.25,
882
+ "step": 71
883
+ },
884
+ {
885
+ "loss": 0.0167,
886
+ "learning_rate": 0.0002,
887
+ "epoch": 1.26,
888
+ "step": 72
889
+ },
890
+ {
891
+ "loss": 0.0088,
892
+ "learning_rate": 0.0002,
893
+ "epoch": 1.28,
894
+ "step": 73
895
+ },
896
+ {
897
+ "loss": 0.0003,
898
+ "learning_rate": 0.0002,
899
+ "epoch": 1.3,
900
+ "step": 74
901
+ },
902
+ {
903
+ "loss": 0.0111,
904
+ "learning_rate": 0.0002,
905
+ "epoch": 1.32,
906
+ "step": 75
907
+ },
908
+ {
909
+ "eval_code_easy_loss": 0.08589933067560196,
910
+ "eval_code_easy_score": -0.029182305559515953,
911
+ "eval_code_easy_brier_score": 0.029182305559515953,
912
+ "eval_code_easy_average_probability": 0.9475240111351013,
913
+ "eval_code_easy_accuracy": 0.95,
914
+ "eval_code_easy_probabilities": [
915
+ 1.0,
916
+ 1.0,
917
+ 1.0,
918
+ 1.0,
919
+ 1.0,
920
+ 0.9999991655349731,
921
+ 0.44722074270248413,
922
+ 1.0,
923
+ 1.0,
924
+ 0.9999874830245972,
925
+ 0.9999817609786987,
926
+ 0.9999996423721313,
927
+ 0.9983553290367126,
928
+ 1.0,
929
+ 0.9999998807907104,
930
+ 0.969874918460846,
931
+ 1.0,
932
+ 1.0,
933
+ 1.0,
934
+ 0.9999077320098877,
935
+ 0.7078168392181396,
936
+ 1.0,
937
+ 1.0,
938
+ 1.0,
939
+ 0.9998465776443481,
940
+ 0.9997827410697937,
941
+ 0.9978830218315125,
942
+ 0.9988357424736023,
943
+ 0.9989570379257202,
944
+ 0.9992606043815613,
945
+ 0.7118145227432251,
946
+ 0.8777998089790344,
947
+ 0.7065884470939636,
948
+ 0.9980650544166565,
949
+ 0.9990893602371216,
950
+ 0.9781619310379028,
951
+ 0.9999847412109375,
952
+ 1.0,
953
+ 0.9993658661842346,
954
+ 0.9999990463256836,
955
+ 0.9999980926513672,
956
+ 0.9999912977218628,
957
+ 0.9999923706054688,
958
+ 1.0,
959
+ 0.9999966621398926,
960
+ 0.9973452687263489,
961
+ 0.13087554275989532,
962
+ 0.8122653365135193,
963
+ 0.9999538660049438,
964
+ 0.9999990463256836,
965
+ 0.9998875856399536,
966
+ 0.9902234673500061,
967
+ 1.0,
968
+ 1.0,
969
+ 0.9999982118606567,
970
+ 0.9999996423721313,
971
+ 0.9999994039535522,
972
+ 0.9999998807907104,
973
+ 0.396033376455307,
974
+ 0.9999998807907104,
975
+ 0.9999998807907104,
976
+ 1.0,
977
+ 0.9975220561027527,
978
+ 1.0,
979
+ 1.0,
980
+ 1.0,
981
+ 0.9999996423721313,
982
+ 1.0,
983
+ 0.9999997615814209,
984
+ 0.9743689298629761,
985
+ 0.3422013521194458,
986
+ 0.25521236658096313,
987
+ 0.9973738193511963,
988
+ 0.9997031092643738,
989
+ 0.9819728136062622,
990
+ 0.9939416646957397,
991
+ 0.997735857963562,
992
+ 0.9992177486419678,
993
+ 1.0,
994
+ 0.9999991655349731,
995
+ 0.9999998807907104,
996
+ 0.9803919196128845,
997
+ 1.0,
998
+ 0.9999996423721313,
999
+ 0.9999098777770996,
1000
+ 0.9999998807907104,
1001
+ 0.9999986886978149,
1002
+ 0.9992156028747559,
1003
+ 0.9822015166282654,
1004
+ 0.9914659261703491,
1005
+ 0.9999808073043823,
1006
+ 0.5561801791191101,
1007
+ 0.9888289570808411,
1008
+ 1.0,
1009
+ 1.0,
1010
+ 1.0,
1011
+ 1.0,
1012
+ 1.0,
1013
+ 1.0,
1014
+ 0.9998490810394287
1015
+ ],
1016
+ "eval_code_easy_runtime": 31.1417,
1017
+ "eval_code_easy_samples_per_second": 3.211,
1018
+ "eval_code_easy_steps_per_second": 0.128,
1019
+ "epoch": 1.32,
1020
+ "step": 75
1021
+ },
1022
+ {
1023
+ "eval_code_hard_loss": 0.4541158378124237,
1024
+ "eval_code_hard_score": -0.15187427401542664,
1025
+ "eval_code_hard_brier_score": 0.15187427401542664,
1026
+ "eval_code_hard_average_probability": 0.7136721014976501,
1027
+ "eval_code_hard_accuracy": 0.79,
1028
+ "eval_code_hard_probabilities": [
1029
+ 0.9999998807907104,
1030
+ 0.42837128043174744,
1031
+ 0.6079556345939636,
1032
+ 0.9998014569282532,
1033
+ 0.9960059523582458,
1034
+ 0.9952560067176819,
1035
+ 0.25668519735336304,
1036
+ 0.37132731080055237,
1037
+ 0.18430009484291077,
1038
+ 0.5929830074310303,
1039
+ 0.3093147277832031,
1040
+ 0.7880420088768005,
1041
+ 0.999980092048645,
1042
+ 0.9994723200798035,
1043
+ 0.9992823004722595,
1044
+ 0.9913004040718079,
1045
+ 0.9134578704833984,
1046
+ 0.9550933837890625,
1047
+ 0.9999967813491821,
1048
+ 0.9999939203262329,
1049
+ 0.9999887943267822,
1050
+ 0.9903974533081055,
1051
+ 0.9929130673408508,
1052
+ 0.9700443744659424,
1053
+ 0.5675995945930481,
1054
+ 0.4452306032180786,
1055
+ 0.39090991020202637,
1056
+ 0.05753504857420921,
1057
+ 0.5650666952133179,
1058
+ 0.6610203981399536,
1059
+ 0.8633843660354614,
1060
+ 0.9552722573280334,
1061
+ 0.6663053035736084,
1062
+ 0.7623808979988098,
1063
+ 0.4418239891529083,
1064
+ 0.5061418414115906,
1065
+ 0.2837110459804535,
1066
+ 0.19440007209777832,
1067
+ 0.4149368107318878,
1068
+ 0.6865063905715942,
1069
+ 0.653810441493988,
1070
+ 0.8122990131378174,
1071
+ 0.7234838604927063,
1072
+ 0.5346352458000183,
1073
+ 0.6917294263839722,
1074
+ 0.9355289340019226,
1075
+ 0.5552300810813904,
1076
+ 0.6327811479568481,
1077
+ 0.8747856020927429,
1078
+ 0.962817907333374,
1079
+ 0.9655711650848389,
1080
+ 0.18094369769096375,
1081
+ 0.8358926177024841,
1082
+ 0.876140296459198,
1083
+ 0.9999440908432007,
1084
+ 0.9999480247497559,
1085
+ 0.9999721050262451,
1086
+ 0.2512325644493103,
1087
+ 0.5562063455581665,
1088
+ 0.623430609703064,
1089
+ 0.9999799728393555,
1090
+ 0.9999500513076782,
1091
+ 0.9999836683273315,
1092
+ 0.9980340600013733,
1093
+ 0.9961856007575989,
1094
+ 0.9999278783798218,
1095
+ 0.9568101167678833,
1096
+ 0.9999703168869019,
1097
+ 0.9811659455299377,
1098
+ 0.9707834720611572,
1099
+ 0.8236635327339172,
1100
+ 0.4347587823867798,
1101
+ 0.8745930790901184,
1102
+ 0.9302994608879089,
1103
+ 0.9327366352081299,
1104
+ 0.766298234462738,
1105
+ 0.3869512379169464,
1106
+ 0.4064312279224396,
1107
+ 0.5269957184791565,
1108
+ 0.5092197060585022,
1109
+ 0.5739120841026306,
1110
+ 0.9161906838417053,
1111
+ 0.9152584075927734,
1112
+ 0.8292853832244873,
1113
+ 0.6077045798301697,
1114
+ 0.4789629280567169,
1115
+ 0.55214923620224,
1116
+ 0.5199791193008423,
1117
+ 0.834286093711853,
1118
+ 0.7272135615348816,
1119
+ 0.5919612050056458,
1120
+ 0.628190279006958,
1121
+ 0.5904282331466675,
1122
+ 0.20717588067054749,
1123
+ 0.23157484829425812,
1124
+ 0.17664135992527008,
1125
+ 0.7453339099884033,
1126
+ 0.9322812557220459,
1127
+ 0.7623251676559448,
1128
+ 0.5810563564300537
1129
+ ],
1130
+ "eval_code_hard_runtime": 110.6654,
1131
+ "eval_code_hard_samples_per_second": 0.904,
1132
+ "eval_code_hard_steps_per_second": 0.036,
1133
+ "epoch": 1.32,
1134
+ "step": 75
1135
+ },
1136
+ {
1137
+ "loss": 0.0328,
1138
+ "learning_rate": 0.0002,
1139
+ "epoch": 1.33,
1140
+ "step": 76
1141
+ },
1142
+ {
1143
+ "loss": 0.0814,
1144
+ "learning_rate": 0.0002,
1145
+ "epoch": 1.35,
1146
+ "step": 77
1147
+ },
1148
+ {
1149
+ "loss": 0.0022,
1150
+ "learning_rate": 0.0002,
1151
+ "epoch": 1.37,
1152
+ "step": 78
1153
+ },
1154
+ {
1155
+ "loss": 0.0048,
1156
+ "learning_rate": 0.0002,
1157
+ "epoch": 1.39,
1158
+ "step": 79
1159
+ },
1160
+ {
1161
+ "loss": 0.0439,
1162
+ "learning_rate": 0.0002,
1163
+ "epoch": 1.4,
1164
+ "step": 80
1165
+ },
1166
+ {
1167
+ "loss": 0.0316,
1168
+ "learning_rate": 0.0002,
1169
+ "epoch": 1.42,
1170
+ "step": 81
1171
+ },
1172
+ {
1173
+ "loss": 0.0956,
1174
+ "learning_rate": 0.0002,
1175
+ "epoch": 1.44,
1176
+ "step": 82
1177
+ },
1178
+ {
1179
+ "loss": 0.0138,
1180
+ "learning_rate": 0.0002,
1181
+ "epoch": 1.46,
1182
+ "step": 83
1183
+ },
1184
+ {
1185
+ "loss": 0.0014,
1186
+ "learning_rate": 0.0002,
1187
+ "epoch": 1.47,
1188
+ "step": 84
1189
+ },
1190
+ {
1191
+ "loss": 0.0353,
1192
+ "learning_rate": 0.0002,
1193
+ "epoch": 1.49,
1194
+ "step": 85
1195
+ },
1196
+ {
1197
+ "loss": 0.0024,
1198
+ "learning_rate": 0.0002,
1199
+ "epoch": 1.51,
1200
+ "step": 86
1201
+ },
1202
+ {
1203
+ "loss": 0.0206,
1204
+ "learning_rate": 0.0002,
1205
+ "epoch": 1.53,
1206
+ "step": 87
1207
+ },
1208
+ {
1209
+ "loss": 0.0007,
1210
+ "learning_rate": 0.0002,
1211
+ "epoch": 1.54,
1212
+ "step": 88
1213
+ },
1214
+ {
1215
+ "loss": 0.0472,
1216
+ "learning_rate": 0.0002,
1217
+ "epoch": 1.56,
1218
+ "step": 89
1219
+ },
1220
+ {
1221
+ "loss": 0.0521,
1222
+ "learning_rate": 0.0002,
1223
+ "epoch": 1.58,
1224
+ "step": 90
1225
+ },
1226
+ {
1227
+ "loss": 0.0073,
1228
+ "learning_rate": 0.0002,
1229
+ "epoch": 1.6,
1230
+ "step": 91
1231
+ },
1232
+ {
1233
+ "loss": 0.1226,
1234
+ "learning_rate": 0.0002,
1235
+ "epoch": 1.61,
1236
+ "step": 92
1237
+ },
1238
+ {
1239
+ "loss": 0.0452,
1240
+ "learning_rate": 0.0002,
1241
+ "epoch": 1.63,
1242
+ "step": 93
1243
+ },
1244
+ {
1245
+ "loss": 0.0143,
1246
+ "learning_rate": 0.0002,
1247
+ "epoch": 1.65,
1248
+ "step": 94
1249
+ },
1250
+ {
1251
+ "loss": 0.0593,
1252
+ "learning_rate": 0.0002,
1253
+ "epoch": 1.67,
1254
+ "step": 95
1255
+ },
1256
+ {
1257
+ "loss": 0.0165,
1258
+ "learning_rate": 0.0002,
1259
+ "epoch": 1.68,
1260
+ "step": 96
1261
+ },
1262
+ {
1263
+ "loss": 0.0016,
1264
+ "learning_rate": 0.0002,
1265
+ "epoch": 1.7,
1266
+ "step": 97
1267
+ },
1268
+ {
1269
+ "loss": 0.0025,
1270
+ "learning_rate": 0.0002,
1271
+ "epoch": 1.72,
1272
+ "step": 98
1273
+ },
1274
+ {
1275
+ "loss": 0.0032,
1276
+ "learning_rate": 0.0002,
1277
+ "epoch": 1.74,
1278
+ "step": 99
1279
+ },
1280
+ {
1281
+ "loss": 0.0065,
1282
+ "learning_rate": 0.0002,
1283
+ "epoch": 1.75,
1284
+ "step": 100
1285
+ },
1286
+ {
1287
+ "eval_code_easy_loss": 0.1302385777235031,
1288
+ "eval_code_easy_score": -0.038255855441093445,
1289
+ "eval_code_easy_brier_score": 0.038255855441093445,
1290
+ "eval_code_easy_average_probability": 0.9474854469299316,
1291
+ "eval_code_easy_accuracy": 0.96,
1292
+ "eval_code_easy_probabilities": [
1293
+ 0.9999997615814209,
1294
+ 1.0,
1295
+ 0.9999998807907104,
1296
+ 1.0,
1297
+ 0.9999998807907104,
1298
+ 0.9999867677688599,
1299
+ 0.14685745537281036,
1300
+ 1.0,
1301
+ 1.0,
1302
+ 0.9999862909317017,
1303
+ 0.9999769926071167,
1304
+ 0.9999992847442627,
1305
+ 0.9967072606086731,
1306
+ 1.0,
1307
+ 0.9999990463256836,
1308
+ 0.9817181825637817,
1309
+ 1.0,
1310
+ 1.0,
1311
+ 1.0,
1312
+ 0.9999679327011108,
1313
+ 0.989346444606781,
1314
+ 1.0,
1315
+ 1.0,
1316
+ 0.9999996423721313,
1317
+ 1.0,
1318
+ 1.0,
1319
+ 0.9997485280036926,
1320
+ 0.9999996423721313,
1321
+ 0.9999998807907104,
1322
+ 0.9999997615814209,
1323
+ 0.7592384815216064,
1324
+ 0.9866325259208679,
1325
+ 0.7520292401313782,
1326
+ 0.9871376752853394,
1327
+ 0.9395173788070679,
1328
+ 0.9484317898750305,
1329
+ 0.99998939037323,
1330
+ 1.0,
1331
+ 0.987211287021637,
1332
+ 0.9999986886978149,
1333
+ 0.9999985694885254,
1334
+ 0.999853253364563,
1335
+ 0.9999595880508423,
1336
+ 0.9999998807907104,
1337
+ 0.9999158382415771,
1338
+ 0.9994738698005676,
1339
+ 0.07437732070684433,
1340
+ 0.8850747346878052,
1341
+ 0.9996064305305481,
1342
+ 0.9999996423721313,
1343
+ 0.9999347925186157,
1344
+ 0.9994233846664429,
1345
+ 1.0,
1346
+ 0.9999998807907104,
1347
+ 1.0,
1348
+ 1.0,
1349
+ 1.0,
1350
+ 1.0,
1351
+ 0.754472017288208,
1352
+ 0.9999998807907104,
1353
+ 0.9999997615814209,
1354
+ 1.0,
1355
+ 0.9901873469352722,
1356
+ 1.0,
1357
+ 1.0,
1358
+ 1.0,
1359
+ 1.0,
1360
+ 1.0,
1361
+ 1.0,
1362
+ 0.9974935054779053,
1363
+ 0.0717310905456543,
1364
+ 0.029978960752487183,
1365
+ 1.0,
1366
+ 1.0,
1367
+ 0.999995231628418,
1368
+ 0.9984604120254517,
1369
+ 0.9997397065162659,
1370
+ 0.9999804496765137,
1371
+ 1.0,
1372
+ 1.0,
1373
+ 1.0,
1374
+ 0.9670922160148621,
1375
+ 0.9999996423721313,
1376
+ 0.9999791383743286,
1377
+ 0.998187243938446,
1378
+ 0.9999996423721313,
1379
+ 0.9999892711639404,
1380
+ 0.9999653100967407,
1381
+ 0.9966863989830017,
1382
+ 0.999233603477478,
1383
+ 1.0,
1384
+ 0.513279378414154,
1385
+ 0.9999966621398926,
1386
+ 1.0,
1387
+ 1.0,
1388
+ 1.0,
1389
+ 1.0,
1390
+ 1.0,
1391
+ 1.0,
1392
+ 0.9999997615814209
1393
+ ],
1394
+ "eval_code_easy_runtime": 31.1546,
1395
+ "eval_code_easy_samples_per_second": 3.21,
1396
+ "eval_code_easy_steps_per_second": 0.128,
1397
+ "epoch": 1.75,
1398
+ "step": 100
1399
+ },
1400
+ {
1401
+ "eval_code_hard_loss": 0.4152863323688507,
1402
+ "eval_code_hard_score": -0.13397492468357086,
1403
+ "eval_code_hard_brier_score": 0.13397492468357086,
1404
+ "eval_code_hard_average_probability": 0.7467895746231079,
1405
+ "eval_code_hard_accuracy": 0.82,
1406
+ "eval_code_hard_probabilities": [
1407
+ 0.999998927116394,
1408
+ 0.3584747314453125,
1409
+ 0.5524832010269165,
1410
+ 0.9999052286148071,
1411
+ 0.9991750121116638,
1412
+ 0.9992142915725708,
1413
+ 0.17623497545719147,
1414
+ 0.41959860920906067,
1415
+ 0.08018570393323898,
1416
+ 0.6593999862670898,
1417
+ 0.26273950934410095,
1418
+ 0.7871806621551514,
1419
+ 0.999994158744812,
1420
+ 0.9992138147354126,
1421
+ 0.9983713030815125,
1422
+ 0.9971001744270325,
1423
+ 0.9297299981117249,
1424
+ 0.9683568477630615,
1425
+ 0.9999955892562866,
1426
+ 0.9999574422836304,
1427
+ 0.9999104738235474,
1428
+ 0.9948668479919434,
1429
+ 0.9949619770050049,
1430
+ 0.9704132080078125,
1431
+ 0.6137145757675171,
1432
+ 0.5000441074371338,
1433
+ 0.39442917704582214,
1434
+ 0.11175673454999924,
1435
+ 0.797296941280365,
1436
+ 0.8252298831939697,
1437
+ 0.9333888292312622,
1438
+ 0.9808279871940613,
1439
+ 0.7203049659729004,
1440
+ 0.8538256287574768,
1441
+ 0.4518895745277405,
1442
+ 0.613561749458313,
1443
+ 0.5261998772621155,
1444
+ 0.38816678524017334,
1445
+ 0.7502338886260986,
1446
+ 0.6585968136787415,
1447
+ 0.6722922325134277,
1448
+ 0.8558183908462524,
1449
+ 0.8481931090354919,
1450
+ 0.5945469737052917,
1451
+ 0.8612274527549744,
1452
+ 0.9837346076965332,
1453
+ 0.5298483371734619,
1454
+ 0.7162337303161621,
1455
+ 0.9290578365325928,
1456
+ 0.9855068922042847,
1457
+ 0.9883608818054199,
1458
+ 0.20835253596305847,
1459
+ 0.9197254776954651,
1460
+ 0.8866751194000244,
1461
+ 0.9997254014015198,
1462
+ 0.9997286200523376,
1463
+ 0.9998807907104492,
1464
+ 0.4348519742488861,
1465
+ 0.5857678651809692,
1466
+ 0.5733372569084167,
1467
+ 0.9998024106025696,
1468
+ 0.9995396137237549,
1469
+ 0.9998968839645386,
1470
+ 0.9989103078842163,
1471
+ 0.9946140646934509,
1472
+ 0.9999970197677612,
1473
+ 0.8636531829833984,
1474
+ 0.9994328618049622,
1475
+ 0.9395280480384827,
1476
+ 0.9852699637413025,
1477
+ 0.8448395729064941,
1478
+ 0.33028489351272583,
1479
+ 0.9747959971427917,
1480
+ 0.996696949005127,
1481
+ 0.9959298968315125,
1482
+ 0.8669978976249695,
1483
+ 0.32030242681503296,
1484
+ 0.336773544549942,
1485
+ 0.5315243005752563,
1486
+ 0.4963895082473755,
1487
+ 0.5617468357086182,
1488
+ 0.9487797021865845,
1489
+ 0.9507840275764465,
1490
+ 0.8413812518119812,
1491
+ 0.8480523228645325,
1492
+ 0.5873783230781555,
1493
+ 0.5967154502868652,
1494
+ 0.5515139698982239,
1495
+ 0.7987790107727051,
1496
+ 0.7132774591445923,
1497
+ 0.7350783348083496,
1498
+ 0.7808082699775696,
1499
+ 0.746001124382019,
1500
+ 0.17566508054733276,
1501
+ 0.21035020053386688,
1502
+ 0.10222747921943665,
1503
+ 0.6762077212333679,
1504
+ 0.9854094386100769,
1505
+ 0.8317046761512756,
1506
+ 0.7221127152442932
1507
+ ],
1508
+ "eval_code_hard_runtime": 110.7468,
1509
+ "eval_code_hard_samples_per_second": 0.903,
1510
+ "eval_code_hard_steps_per_second": 0.036,
1511
+ "epoch": 1.75,
1512
+ "step": 100
1513
+ },
1514
+ {
1515
+ "loss": 0.0416,
1516
+ "learning_rate": 0.0002,
1517
+ "epoch": 1.77,
1518
+ "step": 101
1519
+ },
1520
+ {
1521
+ "loss": 0.0034,
1522
+ "learning_rate": 0.0002,
1523
+ "epoch": 1.79,
1524
+ "step": 102
1525
+ },
1526
+ {
1527
+ "loss": 0.1193,
1528
+ "learning_rate": 0.0002,
1529
+ "epoch": 1.81,
1530
+ "step": 103
1531
+ },
1532
+ {
1533
+ "loss": 0.0152,
1534
+ "learning_rate": 0.0002,
1535
+ "epoch": 1.82,
1536
+ "step": 104
1537
+ },
1538
+ {
1539
+ "loss": 0.0074,
1540
+ "learning_rate": 0.0002,
1541
+ "epoch": 1.84,
1542
+ "step": 105
1543
+ },
1544
+ {
1545
+ "loss": 0.0236,
1546
+ "learning_rate": 0.0002,
1547
+ "epoch": 1.86,
1548
+ "step": 106
1549
+ },
1550
+ {
1551
+ "loss": 0.0077,
1552
+ "learning_rate": 0.0002,
1553
+ "epoch": 1.88,
1554
+ "step": 107
1555
+ },
1556
+ {
1557
+ "loss": 0.1086,
1558
+ "learning_rate": 0.0002,
1559
+ "epoch": 1.89,
1560
+ "step": 108
1561
+ },
1562
+ {
1563
+ "loss": 0.0718,
1564
+ "learning_rate": 0.0002,
1565
+ "epoch": 1.91,
1566
+ "step": 109
1567
+ },
1568
+ {
1569
+ "loss": 0.0042,
1570
+ "learning_rate": 0.0002,
1571
+ "epoch": 1.93,
1572
+ "step": 110
1573
+ },
1574
+ {
1575
+ "loss": 0.0068,
1576
+ "learning_rate": 0.0002,
1577
+ "epoch": 1.95,
1578
+ "step": 111
1579
+ },
1580
+ {
1581
+ "loss": 0.0774,
1582
+ "learning_rate": 0.0002,
1583
+ "epoch": 1.96,
1584
+ "step": 112
1585
+ },
1586
+ {
1587
+ "loss": 0.0374,
1588
+ "learning_rate": 0.0002,
1589
+ "epoch": 1.98,
1590
+ "step": 113
1591
+ },
1592
+ {
1593
+ "loss": 0.0035,
1594
+ "learning_rate": 0.0002,
1595
+ "epoch": 2.0,
1596
+ "step": 114
1597
+ },
1598
+ {
1599
+ "loss": 0.0083,
1600
+ "learning_rate": 0.0002,
1601
+ "epoch": 2.02,
1602
+ "step": 115
1603
+ },
1604
+ {
1605
+ "loss": 0.0147,
1606
+ "learning_rate": 0.0002,
1607
+ "epoch": 2.04,
1608
+ "step": 116
1609
+ },
1610
+ {
1611
+ "loss": 0.0079,
1612
+ "learning_rate": 0.0002,
1613
+ "epoch": 2.05,
1614
+ "step": 117
1615
+ },
1616
+ {
1617
+ "loss": 0.0012,
1618
+ "learning_rate": 0.0002,
1619
+ "epoch": 2.07,
1620
+ "step": 118
1621
+ },
1622
+ {
1623
+ "loss": 0.0137,
1624
+ "learning_rate": 0.0002,
1625
+ "epoch": 2.09,
1626
+ "step": 119
1627
+ },
1628
+ {
1629
+ "loss": 0.029,
1630
+ "learning_rate": 0.0002,
1631
+ "epoch": 2.11,
1632
+ "step": 120
1633
+ },
1634
+ {
1635
+ "loss": 0.0125,
1636
+ "learning_rate": 0.0002,
1637
+ "epoch": 2.12,
1638
+ "step": 121
1639
+ },
1640
+ {
1641
+ "loss": 0.0102,
1642
+ "learning_rate": 0.0002,
1643
+ "epoch": 2.14,
1644
+ "step": 122
1645
+ },
1646
+ {
1647
+ "loss": 0.0057,
1648
+ "learning_rate": 0.0002,
1649
+ "epoch": 2.16,
1650
+ "step": 123
1651
+ },
1652
+ {
1653
+ "loss": 0.0001,
1654
+ "learning_rate": 0.0002,
1655
+ "epoch": 2.18,
1656
+ "step": 124
1657
+ },
1658
+ {
1659
+ "loss": 0.1814,
1660
+ "learning_rate": 0.0002,
1661
+ "epoch": 2.19,
1662
+ "step": 125
1663
+ },
1664
+ {
1665
+ "eval_code_easy_loss": 0.10904992371797562,
1666
+ "eval_code_easy_score": -0.03530557453632355,
1667
+ "eval_code_easy_brier_score": 0.03530557453632355,
1668
+ "eval_code_easy_average_probability": 0.9442441463470459,
1669
+ "eval_code_easy_accuracy": 0.95,
1670
+ "eval_code_easy_probabilities": [
1671
+ 0.9999860525131226,
1672
+ 0.9999575614929199,
1673
+ 0.9999876022338867,
1674
+ 1.0,
1675
+ 1.0,
1676
+ 0.9999998807907104,
1677
+ 0.3663080930709839,
1678
+ 1.0,
1679
+ 1.0,
1680
+ 0.9986967444419861,
1681
+ 0.9995228052139282,
1682
+ 0.9994437098503113,
1683
+ 0.9989981055259705,
1684
+ 1.0,
1685
+ 1.0,
1686
+ 0.9903301000595093,
1687
+ 1.0,
1688
+ 1.0,
1689
+ 1.0,
1690
+ 1.0,
1691
+ 0.9906626343727112,
1692
+ 1.0,
1693
+ 1.0,
1694
+ 1.0,
1695
+ 0.9999998807907104,
1696
+ 1.0,
1697
+ 0.9999880790710449,
1698
+ 0.9878363013267517,
1699
+ 0.9968984127044678,
1700
+ 0.9852358102798462,
1701
+ 0.7452288866043091,
1702
+ 0.8594924807548523,
1703
+ 0.625163733959198,
1704
+ 0.9999972581863403,
1705
+ 0.9998786449432373,
1706
+ 0.9998183846473694,
1707
+ 0.9999641180038452,
1708
+ 0.9999997615814209,
1709
+ 0.986954391002655,
1710
+ 0.9998384714126587,
1711
+ 0.9996883869171143,
1712
+ 0.9997907280921936,
1713
+ 1.0,
1714
+ 1.0,
1715
+ 1.0,
1716
+ 0.6630300283432007,
1717
+ 0.3178897202014923,
1718
+ 0.23424135148525238,
1719
+ 0.9830185174942017,
1720
+ 0.9976402521133423,
1721
+ 0.9478833079338074,
1722
+ 1.0,
1723
+ 1.0,
1724
+ 1.0,
1725
+ 1.0,
1726
+ 1.0,
1727
+ 1.0,
1728
+ 0.9999997615814209,
1729
+ 0.9950571060180664,
1730
+ 0.9999996423721313,
1731
+ 0.9999940395355225,
1732
+ 0.9999960660934448,
1733
+ 0.998701810836792,
1734
+ 1.0,
1735
+ 1.0,
1736
+ 1.0,
1737
+ 1.0,
1738
+ 1.0,
1739
+ 1.0,
1740
+ 0.9900442361831665,
1741
+ 0.11046496033668518,
1742
+ 0.05695008859038353,
1743
+ 0.9999970197677612,
1744
+ 0.9999996423721313,
1745
+ 0.999971866607666,
1746
+ 0.9758032560348511,
1747
+ 0.987618088722229,
1748
+ 0.9728375673294067,
1749
+ 1.0,
1750
+ 0.9999998807907104,
1751
+ 1.0,
1752
+ 0.9999992847442627,
1753
+ 1.0,
1754
+ 1.0,
1755
+ 0.9805415868759155,
1756
+ 0.9995923638343811,
1757
+ 0.9991359114646912,
1758
+ 0.9988665580749512,
1759
+ 0.9318742156028748,
1760
+ 0.9750478863716125,
1761
+ 0.9999997615814209,
1762
+ 0.7831814289093018,
1763
+ 0.9958741068840027,
1764
+ 1.0,
1765
+ 1.0,
1766
+ 1.0,
1767
+ 1.0,
1768
+ 1.0,
1769
+ 1.0,
1770
+ 0.9994946718215942
1771
+ ],
1772
+ "eval_code_easy_runtime": 31.118,
1773
+ "eval_code_easy_samples_per_second": 3.214,
1774
+ "eval_code_easy_steps_per_second": 0.129,
1775
+ "epoch": 2.19,
1776
+ "step": 125
1777
+ },
1778
+ {
1779
+ "eval_code_hard_loss": 0.4987069070339203,
1780
+ "eval_code_hard_score": -0.15576696395874023,
1781
+ "eval_code_hard_brier_score": 0.15576696395874023,
1782
+ "eval_code_hard_average_probability": 0.6946533918380737,
1783
+ "eval_code_hard_accuracy": 0.79,
1784
+ "eval_code_hard_probabilities": [
1785
+ 0.9995853304862976,
1786
+ 0.20991425216197968,
1787
+ 0.549527108669281,
1788
+ 1.0,
1789
+ 0.9999998807907104,
1790
+ 1.0,
1791
+ 0.22219569981098175,
1792
+ 0.37956300377845764,
1793
+ 0.20069900155067444,
1794
+ 0.6470227837562561,
1795
+ 0.31780990958213806,
1796
+ 0.6012275218963623,
1797
+ 1.0,
1798
+ 0.9999998807907104,
1799
+ 0.9999998807907104,
1800
+ 0.9296144247055054,
1801
+ 0.8540101051330566,
1802
+ 0.8961200714111328,
1803
+ 1.0,
1804
+ 1.0,
1805
+ 1.0,
1806
+ 0.9873550534248352,
1807
+ 0.9724990129470825,
1808
+ 0.9760501980781555,
1809
+ 0.38157281279563904,
1810
+ 0.42047223448753357,
1811
+ 0.3564426898956299,
1812
+ 0.0027432094793766737,
1813
+ 0.2510468661785126,
1814
+ 0.3818874955177307,
1815
+ 0.8423392176628113,
1816
+ 0.9236337542533875,
1817
+ 0.5938029289245605,
1818
+ 0.8878844380378723,
1819
+ 0.7184504866600037,
1820
+ 0.7954170107841492,
1821
+ 0.4380953311920166,
1822
+ 0.42004096508026123,
1823
+ 0.6841737627983093,
1824
+ 0.5444896817207336,
1825
+ 0.40916869044303894,
1826
+ 0.5416669249534607,
1827
+ 0.7608702182769775,
1828
+ 0.5757149457931519,
1829
+ 0.8424829840660095,
1830
+ 0.9364664554595947,
1831
+ 0.548158586025238,
1832
+ 0.6652321219444275,
1833
+ 0.8666991591453552,
1834
+ 0.876342236995697,
1835
+ 0.8881990313529968,
1836
+ 0.3995114266872406,
1837
+ 0.8002880215644836,
1838
+ 0.6442937254905701,
1839
+ 1.0,
1840
+ 1.0,
1841
+ 1.0,
1842
+ 0.202246755361557,
1843
+ 0.5754519104957581,
1844
+ 0.5530020594596863,
1845
+ 1.0,
1846
+ 1.0,
1847
+ 1.0,
1848
+ 0.9787330031394958,
1849
+ 0.9658524394035339,
1850
+ 0.9971476197242737,
1851
+ 0.7014413475990295,
1852
+ 0.9961981177330017,
1853
+ 0.8666495680809021,
1854
+ 0.8775137066841125,
1855
+ 0.5972762107849121,
1856
+ 0.5123103857040405,
1857
+ 0.6785091757774353,
1858
+ 0.7748872637748718,
1859
+ 0.8229066133499146,
1860
+ 0.6105999946594238,
1861
+ 0.33387431502342224,
1862
+ 0.2667674720287323,
1863
+ 0.520160436630249,
1864
+ 0.5227006673812866,
1865
+ 0.5882536172866821,
1866
+ 0.7376717329025269,
1867
+ 0.8348681926727295,
1868
+ 0.6894320249557495,
1869
+ 0.39057934284210205,
1870
+ 0.3577791750431061,
1871
+ 0.5273920893669128,
1872
+ 0.47946465015411377,
1873
+ 0.7557993531227112,
1874
+ 0.7129892110824585,
1875
+ 0.5154388546943665,
1876
+ 0.55854332447052,
1877
+ 0.5196318030357361,
1878
+ 0.7760440111160278,
1879
+ 0.8043990731239319,
1880
+ 0.7464650869369507,
1881
+ 0.760195791721344,
1882
+ 0.9593195915222168,
1883
+ 0.7079413533210754,
1884
+ 0.5481259822845459
1885
+ ],
1886
+ "eval_code_hard_runtime": 110.6367,
1887
+ "eval_code_hard_samples_per_second": 0.904,
1888
+ "eval_code_hard_steps_per_second": 0.036,
1889
+ "epoch": 2.19,
1890
+ "step": 125
1891
+ },
1892
+ {
1893
+ "loss": 0.0089,
1894
+ "learning_rate": 0.0002,
1895
+ "epoch": 2.21,
1896
+ "step": 126
1897
+ },
1898
+ {
1899
+ "loss": 0.008,
1900
+ "learning_rate": 0.0002,
1901
+ "epoch": 2.23,
1902
+ "step": 127
1903
+ },
1904
+ {
1905
+ "loss": 0.0025,
1906
+ "learning_rate": 0.0002,
1907
+ "epoch": 2.25,
1908
+ "step": 128
1909
+ },
1910
+ {
1911
+ "loss": 0.003,
1912
+ "learning_rate": 0.0002,
1913
+ "epoch": 2.26,
1914
+ "step": 129
1915
+ },
1916
+ {
1917
+ "loss": 0.0029,
1918
+ "learning_rate": 0.0002,
1919
+ "epoch": 2.28,
1920
+ "step": 130
1921
+ },
1922
+ {
1923
+ "loss": 0.012,
1924
+ "learning_rate": 0.0002,
1925
+ "epoch": 2.3,
1926
+ "step": 131
1927
+ },
1928
+ {
1929
+ "loss": 0.0097,
1930
+ "learning_rate": 0.0002,
1931
+ "epoch": 2.32,
1932
+ "step": 132
1933
+ },
1934
+ {
1935
+ "loss": 0.0024,
1936
+ "learning_rate": 0.0002,
1937
+ "epoch": 2.33,
1938
+ "step": 133
1939
+ },
1940
+ {
1941
+ "loss": 0.0047,
1942
+ "learning_rate": 0.0002,
1943
+ "epoch": 2.35,
1944
+ "step": 134
1945
+ },
1946
+ {
1947
+ "loss": 0.0007,
1948
+ "learning_rate": 0.0002,
1949
+ "epoch": 2.37,
1950
+ "step": 135
1951
+ },
1952
+ {
1953
+ "loss": 0.0027,
1954
+ "learning_rate": 0.0002,
1955
+ "epoch": 2.39,
1956
+ "step": 136
1957
+ },
1958
+ {
1959
+ "loss": 0.0,
1960
+ "learning_rate": 0.0002,
1961
+ "epoch": 2.4,
1962
+ "step": 137
1963
+ },
1964
+ {
1965
+ "loss": 0.0001,
1966
+ "learning_rate": 0.0002,
1967
+ "epoch": 2.42,
1968
+ "step": 138
1969
+ },
1970
+ {
1971
+ "loss": 0.0,
1972
+ "learning_rate": 0.0002,
1973
+ "epoch": 2.44,
1974
+ "step": 139
1975
+ },
1976
+ {
1977
+ "loss": 0.0016,
1978
+ "learning_rate": 0.0002,
1979
+ "epoch": 2.46,
1980
+ "step": 140
1981
+ },
1982
+ {
1983
+ "loss": 0.0002,
1984
+ "learning_rate": 0.0002,
1985
+ "epoch": 2.47,
1986
+ "step": 141
1987
+ },
1988
+ {
1989
+ "loss": 0.0696,
1990
+ "learning_rate": 0.0002,
1991
+ "epoch": 2.49,
1992
+ "step": 142
1993
+ },
1994
+ {
1995
+ "loss": 0.0059,
1996
+ "learning_rate": 0.0002,
1997
+ "epoch": 2.51,
1998
+ "step": 143
1999
+ },
2000
+ {
2001
+ "loss": 0.0005,
2002
+ "learning_rate": 0.0002,
2003
+ "epoch": 2.53,
2004
+ "step": 144
2005
+ },
2006
+ {
2007
+ "loss": 0.0,
2008
+ "learning_rate": 0.0002,
2009
+ "epoch": 2.54,
2010
+ "step": 145
2011
+ },
2012
+ {
2013
+ "loss": 0.0204,
2014
+ "learning_rate": 0.0002,
2015
+ "epoch": 2.56,
2016
+ "step": 146
2017
+ },
2018
+ {
2019
+ "loss": 0.0043,
2020
+ "learning_rate": 0.0002,
2021
+ "epoch": 2.58,
2022
+ "step": 147
2023
+ },
2024
+ {
2025
+ "loss": 0.008,
2026
+ "learning_rate": 0.0002,
2027
+ "epoch": 2.6,
2028
+ "step": 148
2029
+ },
2030
+ {
2031
+ "loss": 0.0462,
2032
+ "learning_rate": 0.0002,
2033
+ "epoch": 2.61,
2034
+ "step": 149
2035
+ },
2036
+ {
2037
+ "loss": 0.0002,
2038
+ "learning_rate": 0.0002,
2039
+ "epoch": 2.63,
2040
+ "step": 150
2041
+ },
2042
+ {
2043
+ "eval_code_easy_loss": 0.07808719575405121,
2044
+ "eval_code_easy_score": -0.024909913539886475,
2045
+ "eval_code_easy_brier_score": 0.024909913539886475,
2046
+ "eval_code_easy_average_probability": 0.9644252061843872,
2047
+ "eval_code_easy_accuracy": 0.96,
2048
+ "eval_code_easy_probabilities": [
2049
+ 1.0,
2050
+ 1.0,
2051
+ 1.0,
2052
+ 1.0,
2053
+ 1.0,
2054
+ 1.0,
2055
+ 0.25757572054862976,
2056
+ 1.0,
2057
+ 1.0,
2058
+ 1.0,
2059
+ 1.0,
2060
+ 1.0,
2061
+ 0.9849016070365906,
2062
+ 1.0,
2063
+ 1.0,
2064
+ 0.9876202940940857,
2065
+ 1.0,
2066
+ 1.0,
2067
+ 1.0,
2068
+ 0.9999861717224121,
2069
+ 0.9844418168067932,
2070
+ 1.0,
2071
+ 1.0,
2072
+ 1.0,
2073
+ 1.0,
2074
+ 1.0,
2075
+ 0.9999815225601196,
2076
+ 0.9999998807907104,
2077
+ 1.0,
2078
+ 0.9999998807907104,
2079
+ 0.9330133199691772,
2080
+ 0.9998921155929565,
2081
+ 0.9747464060783386,
2082
+ 0.9998412132263184,
2083
+ 0.9905625581741333,
2084
+ 0.974348783493042,
2085
+ 0.9999979734420776,
2086
+ 1.0,
2087
+ 0.9953662157058716,
2088
+ 1.0,
2089
+ 1.0,
2090
+ 0.9999998807907104,
2091
+ 0.9999034404754639,
2092
+ 1.0,
2093
+ 0.999864935874939,
2094
+ 0.9999926090240479,
2095
+ 0.4541763961315155,
2096
+ 0.9853015542030334,
2097
+ 0.9999998807907104,
2098
+ 1.0,
2099
+ 1.0,
2100
+ 0.9991912245750427,
2101
+ 1.0,
2102
+ 1.0,
2103
+ 1.0,
2104
+ 1.0,
2105
+ 1.0,
2106
+ 1.0,
2107
+ 0.8326194882392883,
2108
+ 1.0,
2109
+ 1.0,
2110
+ 1.0,
2111
+ 0.9997536540031433,
2112
+ 1.0,
2113
+ 1.0,
2114
+ 1.0,
2115
+ 1.0,
2116
+ 1.0,
2117
+ 1.0,
2118
+ 0.9999920129776001,
2119
+ 0.16892309486865997,
2120
+ 0.04870113357901573,
2121
+ 1.0,
2122
+ 1.0,
2123
+ 1.0,
2124
+ 0.9999966621398926,
2125
+ 0.9999991655349731,
2126
+ 1.0,
2127
+ 1.0,
2128
+ 1.0,
2129
+ 1.0,
2130
+ 0.9763332009315491,
2131
+ 1.0,
2132
+ 1.0,
2133
+ 0.9999935626983643,
2134
+ 1.0,
2135
+ 1.0,
2136
+ 1.0,
2137
+ 0.9999570846557617,
2138
+ 0.9999858140945435,
2139
+ 1.0,
2140
+ 0.8955685496330261,
2141
+ 0.9999994039535522,
2142
+ 1.0,
2143
+ 1.0,
2144
+ 1.0,
2145
+ 1.0,
2146
+ 1.0,
2147
+ 1.0,
2148
+ 1.0
2149
+ ],
2150
+ "eval_code_easy_runtime": 31.1286,
2151
+ "eval_code_easy_samples_per_second": 3.212,
2152
+ "eval_code_easy_steps_per_second": 0.128,
2153
+ "epoch": 2.63,
2154
+ "step": 150
2155
+ },
2156
+ {
2157
+ "eval_code_hard_loss": 0.5192895531654358,
2158
+ "eval_code_hard_score": -0.15126831829547882,
2159
+ "eval_code_hard_brier_score": 0.15126831829547882,
2160
+ "eval_code_hard_average_probability": 0.747479259967804,
2161
+ "eval_code_hard_accuracy": 0.8,
2162
+ "eval_code_hard_probabilities": [
2163
+ 1.0,
2164
+ 0.23972108960151672,
2165
+ 0.34699881076812744,
2166
+ 0.9999767541885376,
2167
+ 0.9995144605636597,
2168
+ 0.9983108043670654,
2169
+ 0.08374932408332825,
2170
+ 0.43437740206718445,
2171
+ 0.01867709495127201,
2172
+ 0.6382063031196594,
2173
+ 0.15230077505111694,
2174
+ 0.8396382331848145,
2175
+ 0.9999982118606567,
2176
+ 0.9995055198669434,
2177
+ 0.9989895224571228,
2178
+ 0.999188244342804,
2179
+ 0.9740042686462402,
2180
+ 0.9889111518859863,
2181
+ 0.9999994039535522,
2182
+ 0.9998956918716431,
2183
+ 0.9998466968536377,
2184
+ 0.9999018907546997,
2185
+ 0.999890923500061,
2186
+ 0.9991710186004639,
2187
+ 0.5053916573524475,
2188
+ 0.39974549412727356,
2189
+ 0.42459815740585327,
2190
+ 0.0008407415589317679,
2191
+ 0.42807716131210327,
2192
+ 0.5588884949684143,
2193
+ 0.9891363382339478,
2194
+ 0.9961511492729187,
2195
+ 0.7021331191062927,
2196
+ 0.9854065179824829,
2197
+ 0.8423302173614502,
2198
+ 0.8964764475822449,
2199
+ 0.1724509447813034,
2200
+ 0.12598082423210144,
2201
+ 0.7715829610824585,
2202
+ 0.7074795365333557,
2203
+ 0.5872290134429932,
2204
+ 0.8601648211479187,
2205
+ 0.9107590317726135,
2206
+ 0.6049399375915527,
2207
+ 0.9440699219703674,
2208
+ 0.9954397082328796,
2209
+ 0.5415297150611877,
2210
+ 0.787742555141449,
2211
+ 0.9602062702178955,
2212
+ 0.9986072182655334,
2213
+ 0.9982185959815979,
2214
+ 0.1667134314775467,
2215
+ 0.9801592826843262,
2216
+ 0.9148990511894226,
2217
+ 0.9998517036437988,
2218
+ 0.9998817443847656,
2219
+ 0.9999765157699585,
2220
+ 0.3098694384098053,
2221
+ 0.6319935917854309,
2222
+ 0.5650399923324585,
2223
+ 0.999991774559021,
2224
+ 0.9999759197235107,
2225
+ 0.999995231628418,
2226
+ 0.9999890327453613,
2227
+ 0.9999120235443115,
2228
+ 1.0,
2229
+ 0.9366697072982788,
2230
+ 0.9999974966049194,
2231
+ 0.9823035597801208,
2232
+ 0.9867276549339294,
2233
+ 0.8337392807006836,
2234
+ 0.13807430863380432,
2235
+ 0.9943860769271851,
2236
+ 0.9992918968200684,
2237
+ 0.9986520409584045,
2238
+ 0.8924828767776489,
2239
+ 0.1779462993144989,
2240
+ 0.14102093875408173,
2241
+ 0.5240833163261414,
2242
+ 0.4705985486507416,
2243
+ 0.5385470390319824,
2244
+ 0.9867573976516724,
2245
+ 0.9934804439544678,
2246
+ 0.9309672713279724,
2247
+ 0.7547566890716553,
2248
+ 0.46812355518341064,
2249
+ 0.6644623279571533,
2250
+ 0.5478729605674744,
2251
+ 0.9080039262771606,
2252
+ 0.9042761325836182,
2253
+ 0.564723789691925,
2254
+ 0.7475408911705017,
2255
+ 0.5996127128601074,
2256
+ 0.6277127861976624,
2257
+ 0.6707883477210999,
2258
+ 0.3496539294719696,
2259
+ 0.9058393836021423,
2260
+ 0.9995261430740356,
2261
+ 0.902912437915802,
2262
+ 0.6317912936210632
2263
+ ],
2264
+ "eval_code_hard_runtime": 110.7028,
2265
+ "eval_code_hard_samples_per_second": 0.903,
2266
+ "eval_code_hard_steps_per_second": 0.036,
2267
+ "epoch": 2.63,
2268
+ "step": 150
2269
+ },
2270
+ {
2271
+ "train_runtime": 4923.5411,
2272
+ "train_samples_per_second": 0.975,
2273
+ "train_steps_per_second": 0.03,
2274
+ "total_flos": 0.0,
2275
+ "train_loss": 0.09649283700817553,
2276
+ "epoch": 2.63,
2277
+ "step": 150
2278
+ }
2279
+ ]